Merge branch 'AUTOMATIC1111:master' into master

This commit is contained in:
papuSpartan 2022-11-06 21:05:28 -06:00 committed by GitHub
commit 00ebc26c4e
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
50 changed files with 2541 additions and 735 deletions

View file

@ -1,13 +1,12 @@
* @AUTOMATIC1111
/localizations/ar_AR.json @xmodar @blackneoo
/localizations/de_DE.json @LunixWasTaken
/localizations/es_ES.json @innovaciones
/localizations/fr_FR.json @tumbly
/localizations/it_IT.json @EugenioBuffo
/localizations/ja_JP.json @yuuki76
/localizations/ko_KR.json @36DB
/localizations/pt_BR.json @M-art-ucci
/localizations/ru_RU.json @kabachuha
/localizations/tr_TR.json @camenduru
/localizations/zh_CN.json @dtlnor @bgluminous
/localizations/zh_TW.json @benlisquare
# if you were managing a localization and were removed from this file, this is because
# the intended way to do localizations now is via extensions. See:
# https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Developing-extensions
# Make a repo with your localization and since you are still listed as a collaborator
# you can add it to the wiki page yourself. This change is because some people complained
# the git commit log is cluttered with things unrelated to almost everyone and
# because I believe this is the best overall for the project to handle localizations almost
# entirely without my oversight.

View file

@ -155,14 +155,15 @@ The documentation was moved from this README over to the project's [wiki](https:
- Swin2SR - https://github.com/mv-lab/swin2sr
- LDSR - https://github.com/Hafiidz/latent-diffusion
- Ideas for optimizations - https://github.com/basujindal/stable-diffusion
- Doggettx - Cross Attention layer optimization - https://github.com/Doggettx/stable-diffusion, original idea for prompt editing.
- InvokeAI, lstein - Cross Attention layer optimization - https://github.com/invoke-ai/InvokeAI (originally http://github.com/lstein/stable-diffusion)
- Rinon Gal - Textual Inversion - https://github.com/rinongal/textual_inversion (we're not using his code, but we are using his ideas).
- Cross Attention layer optimization - Doggettx - https://github.com/Doggettx/stable-diffusion, original idea for prompt editing.
- Cross Attention layer optimization - InvokeAI, lstein - https://github.com/invoke-ai/InvokeAI (originally http://github.com/lstein/stable-diffusion)
- Textual Inversion - Rinon Gal - https://github.com/rinongal/textual_inversion (we're not using his code, but we are using his ideas).
- Idea for SD upscale - https://github.com/jquesnelle/txt2imghd
- Noise generation for outpainting mk2 - https://github.com/parlance-zz/g-diffuser-bot
- CLIP interrogator idea and borrowing some code - https://github.com/pharmapsychotic/clip-interrogator
- Idea for Composable Diffusion - https://github.com/energy-based-model/Compositional-Visual-Generation-with-Composable-Diffusion-Models-PyTorch
- xformers - https://github.com/facebookresearch/xformers
- DeepDanbooru - interrogator for anime diffusers https://github.com/KichangKim/DeepDanbooru
- Security advice - RyotaK
- Initial Gradio script - posted on 4chan by an Anonymous user. Thank you Anonymous user.
- (You)

View file

@ -1,7 +1,6 @@
addEventListener('keydown', (event) => {
let target = event.originalTarget || event.composedPath()[0];
if (!target.hasAttribute("placeholder")) return;
if (!target.placeholder.toLowerCase().includes("prompt")) return;
if (!target.matches("#toprow textarea.gr-text-input[placeholder]")) return;
if (! (event.metaKey || event.ctrlKey)) return;

View file

@ -3,8 +3,21 @@ global_progressbars = {}
galleries = {}
galleryObservers = {}
// this tracks laumnches of window.setTimeout for progressbar to prevent starting a new timeout when the previous is still running
timeoutIds = {}
function check_progressbar(id_part, id_progressbar, id_progressbar_span, id_skip, id_interrupt, id_preview, id_gallery){
var progressbar = gradioApp().getElementById(id_progressbar)
// gradio 3.8's enlightened approach allows them to create two nested div elements inside each other with same id
// every time you use gr.HTML(elem_id='xxx'), so we handle this here
var progressbar = gradioApp().querySelector("#"+id_progressbar+" #"+id_progressbar)
var progressbarParent
if(progressbar){
progressbarParent = gradioApp().querySelector("#"+id_progressbar)
} else{
progressbar = gradioApp().getElementById(id_progressbar)
progressbarParent = null
}
var skip = id_skip ? gradioApp().getElementById(id_skip) : null
var interrupt = gradioApp().getElementById(id_interrupt)
@ -26,18 +39,26 @@ function check_progressbar(id_part, id_progressbar, id_progressbar_span, id_skip
global_progressbars[id_progressbar] = progressbar
var mutationObserver = new MutationObserver(function(m){
if(timeoutIds[id_part]) return;
preview = gradioApp().getElementById(id_preview)
gallery = gradioApp().getElementById(id_gallery)
if(preview != null && gallery != null){
preview.style.width = gallery.clientWidth + "px"
preview.style.height = gallery.clientHeight + "px"
if(progressbarParent) progressbar.style.width = progressbarParent.clientWidth + "px"
//only watch gallery if there is a generation process going on
check_gallery(id_gallery);
var progressDiv = gradioApp().querySelectorAll('#' + id_progressbar_span).length > 0;
if(!progressDiv){
if(progressDiv){
timeoutIds[id_part] = window.setTimeout(function() {
timeoutIds[id_part] = null
requestMoreProgress(id_part, id_progressbar_span, id_skip, id_interrupt)
}, 500)
} else{
if (skip) {
skip.style.display = "none"
}
@ -47,13 +68,10 @@ function check_progressbar(id_part, id_progressbar, id_progressbar_span, id_skip
if (galleryObservers[id_gallery]) {
galleryObservers[id_gallery].disconnect();
galleries[id_gallery] = null;
}
}
}
}
window.setTimeout(function() { requestMoreProgress(id_part, id_progressbar_span, id_skip, id_interrupt) }, 500)
});
mutationObserver.observe( progressbar, { childList:true, subtree:true })
}

View file

@ -248,4 +248,6 @@ function update_token_counter(button_id) {
function restart_reload(){
document.body.innerHTML='<h1 style="font-family:monospace;margin-top:20%;color:lightgray;text-align:center;">Reloading...</h1>';
setTimeout(function(){location.reload()},2000)
return []
}

View file

@ -142,7 +142,7 @@ def prepare_enviroment():
stable_diffusion_commit_hash = os.environ.get('STABLE_DIFFUSION_COMMIT_HASH', "69ae4b35e0a0f6ee1af8bb9a5d0016ccb27e36dc")
taming_transformers_commit_hash = os.environ.get('TAMING_TRANSFORMERS_COMMIT_HASH', "24268930bf1dce879235a7fddd0b2355b84d7ea6")
k_diffusion_commit_hash = os.environ.get('K_DIFFUSION_COMMIT_HASH', "f4e99857772fc3a126ba886aadf795a332774878")
k_diffusion_commit_hash = os.environ.get('K_DIFFUSION_COMMIT_HASH', "60e5042ca0da89c14d1dd59d73883280f8fce991")
codeformer_commit_hash = os.environ.get('CODEFORMER_COMMIT_HASH', "c5b4593074ba6214284d6acd5f1719b6c5d739af")
blip_commit_hash = os.environ.get('BLIP_COMMIT_HASH', "48211a1594f1321b00f14c9f7a5b4813144b2fb9")
@ -238,12 +238,15 @@ def tests(argv):
proc.kill()
def start_webui():
print(f"Launching Web UI with arguments: {' '.join(sys.argv[1:])}")
def start():
print(f"Launching {'API server' if '--nowebui' in sys.argv else 'Web UI'} with arguments: {' '.join(sys.argv[1:])}")
import webui
webui.webui()
if '--nowebui' in sys.argv:
webui.api_only()
else:
webui.webui()
if __name__ == "__main__":
prepare_enviroment()
start_webui()
start()

View file

@ -70,7 +70,7 @@
"None": "Nichts",
"Prompt matrix": "Promptmatrix",
"Prompts from file or textbox": "Prompts aus Datei oder Textfeld",
"X/Y plot": "X/Y Graf",
"X/Y plot": "X/Y Graph",
"Put variable parts at start of prompt": "Variable teile am start des Prompt setzen",
"Iterate seed every line": "Iterate seed every line",
"List of prompt inputs": "List of prompt inputs",
@ -455,4 +455,4 @@
"Only applies to inpainting models. Determines how strongly to mask off the original image for inpainting and img2img. 1.0 means fully masked, which is the default behaviour. 0.0 means a fully unmasked conditioning. Lower values will help preserve the overall composition of the image, but will struggle with large changes.": "Gilt nur für Inpainting-Modelle. Legt fest, wie stark das Originalbild für Inpainting und img2img maskiert werden soll. 1.0 bedeutet vollständig maskiert, was das Standardverhalten ist. 0.0 bedeutet eine vollständig unmaskierte Konditionierung. Niedrigere Werte tragen dazu bei, die Gesamtkomposition des Bildes zu erhalten, sind aber bei großen Änderungen problematisch.",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "Liste von Einstellungsnamen, getrennt durch Kommas, für Einstellungen, die in der Schnellzugriffsleiste oben erscheinen sollen, anstatt in dem üblichen Einstellungs-Tab. Siehe modules/shared.py für Einstellungsnamen. Erfordert einen Neustart zur Anwendung.",
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "Wenn dieser Wert ungleich Null ist, wird er zum Seed addiert und zur Initialisierung des RNG für Noise bei der Verwendung von Samplern mit Eta verwendet. Dies kann verwendet werden, um noch mehr Variationen von Bildern zu erzeugen, oder um Bilder von anderer Software zu erzeugen, wenn Sie wissen, was Sie tun."
}
}

View file

@ -5,18 +5,21 @@
"": "",
"": "",
"Loading...": "Cargando...",
"view": "mostrar",
"api": "api",
"•": "",
"built with gradio": "Construido con Gradio",
"view": "Mostrar ",
"api": "API",
"•": "",
"Construido con Gradio": "Construido con Gradio",
"Stable Diffusion checkpoint": "Stable Diffusion checkpoint",
"txt2img": "txt2img",
"img2img": "img2img",
"Extras": "Extras",
"PNG Info": "PNG Info",
"Checkpoint Merger": "Checkpoint Merger",
"Train": "Train",
"PNG Info": "Info PNG",
"Checkpoint Merger": "Fusionar Checkpoints",
"Train": "Entrenar",
"Deforum": "Deforum",
"Image Browser": "Navegador de Imágenes",
"Settings": "Ajustes",
"Extensions": "Extensiones",
"Prompt": "Prompt",
"Negative prompt": "Prompt negativo",
"Run": "Ejecutar",
@ -27,15 +30,15 @@
"Style 2": "Estilo 2",
"Label": "Etiqueta",
"File": "Archivo",
"Drop File Here": "Suelta el Archivo Aquí",
"Coloque el archivo aquí": "Suelta el archivo aquí",
"-": "-",
"o": "o",
"Click to Upload": "Click para Cargar",
"Haga click para cargar": "Haz click para cargar",
"Image": "Imagen",
"Check progress": "Comprobar progreso",
"Check progress (first)": "Comprobar progreso (inicial)",
"Sampling Steps": "Sampling Steps",
"Sampling method": "Sampling method",
"Sampling method": "Método de Sampling",
"Euler a": "Euler a",
"Euler": "Euler",
"LMS": "LMS",
@ -57,33 +60,34 @@
"Firstpass width": "Ancho original",
"Firstpass height": "Alto original",
"Denoising strength": "Denoising strength",
"Batch count": "Batch count",
"Batch size": "Batch size",
"CFG Scale": "CFG Scale",
"Batch count": "Cantidad del Batch",
"Batch size": "Tamaño del Batch",
"CFG Scale": "Escala CFG",
"Seed": "Seed",
"Extra": "Extra",
"Variation seed": "Variation seed",
"Variation strength": "Variation strength",
"Variation seed": "Seed de variación",
"Variation strength": "Fuerza de variación",
"Resize seed from width": "Redimensionar seed del ancho",
"Resize seed from height": "Redimensionar seed del alto",
"Script": "Script",
"None": "Ninguno",
"Prompt matrix": "Prompt matrix",
"Prompt matrix": "Prompt en matriz",
"Prompts from file or textbox": "Prompts desde archivo o campo de texto",
"X/Y plot": "X/Y plot",
"X/Y plot": "Tabla X/Y",
"Put variable parts at start of prompt": "Poner partes variables al inicio del prompt",
"Show Textbox": "Mostrar Campo de texto",
"File with inputs": "Archivo con inputs",
"Prompts": "Prompts",
"X type": "X type",
"Iterate seed every line": "Repetir seed en cada línea",
"Use same random seed for all lines": "Utiliza el mismo seed aleatorio para todas las líneas",
"List of prompt inputs": "Listado de prompts",
"Upload prompt inputs": "Cargar archivo de prompts",
"X type": "X",
"Nothing": "Nada",
"Var. seed": "Var. seed",
"Var. strength": "Var. strength",
"Var. strength": "Var. fuerza",
"Steps": "Steps",
"Prompt S/R": "Prompt S/R",
"Prompt order": "Prompt order",
"Sampler": "Sampler",
"Checkpoint name": "Checkpoint name",
"Checkpoint name": "Nombre Checkpoint",
"Hypernetwork": "Hypernetwork",
"Hypernet str.": "Hypernet str.",
"Sigma Churn": "Sigma Churn",
@ -93,13 +97,13 @@
"Eta": "Eta",
"Clip skip": "Clip skip",
"Denoising": "Denoising",
"X values": "X values",
"Y type": "Y type",
"Y values": "Y values",
"Cond. Image Mask Weight": "Cond. Image Mask Weight",
"X values": "Valores X",
"Y type": "Y",
"Y values": "Valores Y",
"Draw legend": "Agregar leyenda",
"Include Separate Images": "Incluir Imágenes Separadas",
"Keep -1 for seeds": "Mantener -1 para seeds",
"Drop Image Here": "Suelta la Imagen Aquí",
"Save": "Guardar",
"Send to img2img": "Enviar a img2img",
"Send to inpaint": "Enviar a inpaint",
@ -110,22 +114,23 @@
"Inpaint": "Inpaint",
"Batch img2img": "Batch img2img",
"Image for img2img": "Imagen para img2img",
"Image for inpainting with mask": "Imagen para inpainting con mask",
"Mask": "Mask",
"Mask blur": "Mask blur",
"Mask mode": "Mask mode",
"Draw mask": "Dibujar mask",
"Upload mask": "Cargar mask",
"Masking mode": "Masking mode",
"Inpaint masked": "Inpaint masked",
"Inpaint not masked": "Inpaint not masked",
"Masked content": "Masked content",
"fill": "fill",
"Coloque la imagen aquí": "Suelta la imagen aquí",
"Image for inpainting with mask": "Imagen para inpainting con máscara",
"Mask": "Máscara",
"Mask blur": "Difuminar máscara",
"Mask mode": "Modo máscara",
"Draw mask": "Dibujar máscara",
"Upload mask": "Cargar máscara",
"Masking mode": "Modo de enmascarado",
"Inpaint masked": "Inpaint con enmascarado",
"Inpaint not masked": "Inpaint sin enmascarado",
"Masked content": "Contenido enmascarado",
"fill": "rellenar",
"original": "original",
"latent noise": "latent noise",
"latent nothing": "latent nothing",
"Inpaint at full resolution": "Inpaint a resolución completa",
"Inpaint at full resolution padding, pixels": "Inpaint a resolución completa padding, pixeles",
"Inpaint at full resolution padding, pixels": "Inpaint a resolución completa con relleno, en pixeles",
"Process images in a directory on the same machine where the server is running.": "Procesa imágenes en un directorio en la misma máquina donde se ejecuta el servidor.",
"Use an empty output directory to save pictures normally instead of writing to the output directory.": "Usa un directorio de salida vacío para guardar imágenes normalmente en lugar de escribir en el directorio de salida.",
"Input directory": "Directorio de entrada",
@ -139,15 +144,16 @@
"Outpainting mk2": "Outpainting mk2",
"Poor man's outpainting": "Poor man's outpainting",
"SD upscale": "SD upscale",
"Deforum-webui (use tab extension instead!)": "Deforum-webui (utiliza la extensión en su lugar!)",
"should be 2 or lower.": "debe ser 2 o menos.",
"Override `Sampling method` to Euler?(this method is built for it)": "Anular `Sampling method` a Euler? (este método está diseñado para ello)",
"Override `prompt` to the same value as `original prompt`?(and `negative prompt`)": "Anular `prompt` al mismo valor `prompt original`? (y `prompt negativo`)",
"Override `Sampling method` to Euler?(this method is built for it)": "Invalidar `Sampling method` a Euler? (este método está diseñado para ello)",
"Override `prompt` to the same value as `original prompt`?(and `negative prompt`)": "Invalidar `prompt` al mismo valor `prompt original`? (y `prompt negativo`)",
"Original prompt": "Prompt original",
"Original negative prompt": "Promp negativo original",
"Override `Sampling Steps` to the same value as `Decode steps`?": "Anular `Sampling Steps` al mismo valor de `Decode steps`?",
"Override `Sampling Steps` to the same value as `Decode steps`?": "Invalidar `Sampling Steps` al mismo valor de `Decode steps`?",
"Decode steps": "Decode steps",
"Override `Denoising strength` to 1?": "Anular `Denoising strength` a 1?",
"Decode CFG scale": "Decode CFG scale",
"Override `Denoising strength` to 1?": "Invalidar `Denoising strength` a 1?",
"Decode CFG scale": "Decodificar escala CFG",
"Randomness": "Aleatoriedad",
"Sigma adjustment for finding noise for image": "Ajuste Sigma para encontrar ruido para la imagen.",
"Loops": "Loops",
@ -165,29 +171,35 @@
"Tile overlap": "Solapar mosaicos",
"Upscaler": "Upscaler",
"Lanczos": "Lanczos",
"Nearest": "Nearest",
"LDSR": "LDSR",
"SwinIR 4x": "SwinIR 4x",
"ESRGAN_4x": "ESRGAN_4x",
"ScuNET GAN": "ScuNET GAN",
"ScuNET PSNR": "ScuNET PSNR",
"ESRGAN_4x": "ESRGAN_4x",
"SwinIR 4x": "SwinIR 4x",
"Deforum v0.5-webui-beta": "Deforum v0.5-webui-beta",
"This script is deprecated. Please use the full Deforum extension instead.": "Este script está obsoleto. Utiliza la extensión completa de Deforum en su lugar.",
"Update instructions:": "Instrucciones para actualizar:",
"github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/README.md": "github.com/deforum-art/deforum-for-automatic1111-webui/blob/automatic1111-webui/README.md",
"discord.gg/deforum": "discord.gg/deforum",
"Single Image": "Imagen Única",
"Batch Process": "Batch Process",
"Batch from Directory": "Batch desde Directorio",
"Source": "Origen",
"Show result images": "Mostrar resultados de imágenes",
"Show result images": "Mostrar imágenes generadas",
"Scale by": "Escalar por",
"Scale to": "Escalar a",
"Resize": "Redimensionar",
"Crop to fit": "Recortar para ajustar",
"Upscaler 2 visibility": "Upscaler 2 visibilidad",
"GFPGAN visibility": "GFPGAN visibilidad",
"CodeFormer visibility": "CodeFormer visibilidad",
"CodeFormer weight (0 = maximum effect, 1 = minimum effect)": "CodeFormer peso (0 = efecto máximo, 1 = efecto mínimo)",
"Open output directory": "Abrir directorio de salida",
"Upscaler 2 visibility": "Visibilidad Upscaler 2",
"GFPGAN visibility": "Visibilidad GFPGAN",
"CodeFormer visibility": "Visibilidad CodeFormer",
"CodeFormer weight (0 = maximum effect, 1 = minimum effect)": "Influencia CodeFormer (0 = efecto máximo, 1 = efecto mínimo)",
"Upscale Before Restoring Faces": "Escalar antes de Restaurar Rostros",
"Send to txt2img": "Enviar a txt2img",
"A merger of the two checkpoints will be generated in your": "Se generará una fusión de los dos checkpoints en su",
"checkpoint": "checkpoint",
"directory.": "directorio.",
"A merger of the two checkpoints will be generated in your": "Se generará una fusión de los dos checkpoints en tu ",
"checkpoint": "directorio ",
"directory.": "de modelos.",
"Primary model (A)": "Modelo primario (A)",
"Secondary model (B)": "Modelo secundario (B)",
"Tertiary model (C)": "Modelo terciario (C)",
@ -215,6 +227,36 @@
"leakyrelu": "leakyrelu",
"elu": "elu",
"swish": "swish",
"tanh": "tanh",
"sigmoid": "sigmoid",
"celu": "celu",
"gelu": "gelu",
"glu": "glu",
"hardshrink": "hardshrink",
"hardsigmoid": "hardsigmoid",
"hardtanh": "hardtanh",
"logsigmoid": "logsigmoid",
"logsoftmax": "logsoftmax",
"mish": "mish",
"prelu": "prelu",
"rrelu": "rrelu",
"relu6": "relu6",
"selu": "selu",
"silu": "silu",
"softmax": "softmax",
"softmax2d": "softmax2d",
"softmin": "softmin",
"softplus": "softplus",
"softshrink": "softshrink",
"softsign": "softsign",
"tanhshrink": "tanhshrink",
"threshold": "threshold",
"Select Layer weights initialization. relu-like - Kaiming, sigmoid-like - Xavier is recommended": "Seleccionar inicialización de modelos de capa. relu-like - Kaiming, sigmoid-like - Xavier es el recomendado",
"Normal": "Normal",
"KaimingUniform": "KaimingUniform",
"KaimingNormal": "KaimingNormal",
"XavierUniform": "XavierUniform",
"XavierNormal": "XavierNormal",
"Add layer normalization": "Agregar normalización de capa",
"Use dropout": "Usar dropout",
"Overwrite Old Hypernetwork": "Sobrescribir Hypernetwork Anterior",
@ -226,11 +268,16 @@
"prepend": "anteponer",
"append": "añadir",
"Create flipped copies": "Crear copias volteadas",
"Split oversized images": "Dividir imágenes grandes",
"Use BLIP for caption": "Usar BLIP para subtítulo",
"Use deepbooru for caption": "Usar deepbooru para subtítulo",
"Split image threshold": "Threshold imagen dividida",
"Split image overlap ratio": "Overlap ratio imagen dividida",
"Split oversized images": "Dividir imágenes muy grandes",
"Auto focal point crop": "Recorte de punto focal automático",
"Use BLIP for caption": "Usar BLIP para leyenda",
"Use deepbooru for caption": "Usar deepbooru para leyenda",
"Split image threshold": "Umbral en imagen dividida",
"Split image overlap ratio": "Relación de superposición en imagen dividida",
"Focal point face weight": "Peso de la cara del punto focal",
"Focal point entropy weight": "Focal point entropy weight",
"Focal point edges weight": "Focal point edges weight",
"Create debug image": "Crear imagen de depuración",
"Preprocess": "Preproceso",
"Train an embedding or Hypernetwork; you must specify a directory with a set of 1:1 ratio images": "Entrenar un embedding o Hypernetwork; debes especificar un directorio con un conjunto de imágenes con una proporción de 1:1 ",
"[wiki]": "[wiki]",
@ -245,13 +292,194 @@
"Save a copy of embedding to log directory every N steps, 0 to disable": "Guarda una copia de embedding en el directorio log cada N pasos, 0 para deshabilitar",
"Save images with embedding in PNG chunks": "Guarda imágenes con embedding en fragmentos PNG",
"Read parameters (prompt, etc...) from txt2img tab when making previews": "Leer parámetros (prompt, etc...) desde la pestaña txt2img al hacer vistas previas",
"Train Hypernetwork": "Train Hypernetwork",
"Train Embedding": "Train Embedding",
"Train Hypernetwork": "Entrenar Hypernetwork",
"Train Embedding": "Entrenar Embedding",
"Info and links": "Información y enlaces",
"▼": "▼",
"Made by deforum.github.io, port for AUTOMATIC1111's webui maintained by kabachuha": "Desarrolado por deforum.github.io, port para AUTOMATIC1111's webui mantenido por kabachuha",
"Original Deforum Github repo github.com/deforum/stable-diffusion": "Deforum Github repo github.com/deforum/stable-diffusion",
"This fork for auto1111's webui github.com/deforum-art/deforum-for-automatic1111-webui": "Fork para auto1111's webui github.com/deforum-art/deforum-for-automatic1111-webui",
"Join the official Deforum Discord discord.gg/deforum to share your creations and suggestions": "Únete al Discord oficial Deforum discord.gg/deforum para compartir tus creaciones y sugerencias",
"User guide for v0.5 docs.google.com/document/d/1pEobUknMFMkn8F5TMsv8qRzamXX_75BShMMXV8IFslI/edit": "Guía de usuario v0.5 docs.google.com/document/d/1pEobUknMFMkn8F5TMsv8qRzamXX_75BShMMXV8IFslI/edit",
"Math keyframing explanation docs.google.com/document/d/1pfW1PwbDIuW0cv-dnuyYj1UzPqe23BlSLTJsqazffXM/edit?usp=sharing": "Explicación de fotogramas matemáticos docs.google.com/document/d/1pfW1PwbDIuW0cv-dnuyYj1UzPqe23BlSLTJsqazffXM/edit?usp=sharing",
"Keyframes": "Keyframes",
"Prompts": "Prompts",
"Init": "Init",
"Video output": "Salida de vídeo",
"Run settings": "Ejecutar ajustes",
"Import settings from file": "Importar ajustes desde archivo",
"Override settings": "Invalidar ajustes",
"Custom settings file": "Archivo de ajustes personalizados",
"Sampling settings": "Ajustes de sampling",
"override_these_with_webui": "override_these_with_webui",
"W": "Ancho",
"H": "Alto",
"seed": "seed",
"sampler": "sampler",
"Enable extras": "Activar extras",
"subseed": "subseed",
"subseed_strength": "subseed_strength",
"steps": "steps",
"ddim_eta": "ddim_eta",
"n_batch": "n_batch",
"make_grid": "make_grid",
"grid_rows": "grid_rows",
"save_settings": "save_settings",
"save_samples": "save_samples",
"display_samples": "display_samples",
"save_sample_per_step": "save_sample_per_step",
"show_sample_per_step": "show_sample_per_step",
"Batch settings": "Ajustes de batch",
"batch_name": "batch_name",
"filename_format": "filename_format",
"seed_behavior": "seed_behavior",
"iter": "iter",
"fixed": "fixed",
"random": "random",
"schedule": "schedule",
"Animation settings": "Ajustes de animación",
"animation_mode": "animation_mode",
"2D": "2D",
"3D": "3D",
"Video Input": "Entrada de Video",
"max_frames": "max_frames",
"border": "border",
"replicate": "replicate",
"wrap": "wrap",
"Motion parameters:": "Parámetros de movimiento:",
"2D and 3D settings": "Ajustes 2D y 3D",
"angle": "angle",
"zoom": "zoom",
"translation_x": "translation_x",
"translation_y": "translation_y",
"3D settings": "Ajustes 3D",
"translation_z": "translation_z",
"rotation_3d_x": "rotation_3d_x",
"rotation_3d_y": "rotation_3d_y",
"rotation_3d_z": "rotation_3d_z",
"Prespective flip — Low VRAM pseudo-3D mode:": "Prespective flip — Modo Low VRAM pseudo-3D:",
"flip_2d_perspective": "flip_2d_perspective",
"perspective_flip_theta": "perspective_flip_theta",
"perspective_flip_phi": "perspective_flip_phi",
"perspective_flip_gamma": "perspective_flip_gamma",
"perspective_flip_fv": "perspective_flip_fv",
"Generation settings:": "Ajustes de generación:",
"noise_schedule": "noise_schedule",
"strength_schedule": "strength_schedule",
"contrast_schedule": "contrast_schedule",
"cfg_scale_schedule": "cfg_scale_schedule",
"3D Fov settings:": "Ajustes 3D Fov:",
"fov_schedule": "fov_schedule",
"near_schedule": "near_schedule",
"far_schedule": "far_schedule",
"To enable seed schedule select seed behavior — 'schedule'": "Para habilitar el seed schedule, selecciona el comportamiento del seed — 'schedule'",
"seed_schedule": "seed_schedule",
"Coherence:": "Coherencia:",
"color_coherence": "color_coherence",
"Match Frame 0 HSV": "Match Frame 0 HSV",
"Match Frame 0 LAB": "Match Frame 0 LAB",
"Match Frame 0 RGB": "Match Frame 0 RGB",
"diffusion_cadence": "diffusion_cadence",
"3D Depth Warping:": "3D Depth Warping:",
"use_depth_warping": "use_depth_warping",
"midas_weight": "midas_weight",
"near_plane": "near_plane",
"far_plane": "far_plane",
"fov": "fov",
"padding_mode": "padding_mode",
"reflection": "reflection",
"zeros": "zeros",
"sampling_mode": "sampling_mode",
"bicubic": "bicubic",
"bilinear": "bilinear",
"nearest": "nearest",
"save_depth_maps": "save_depth_maps",
"`animation_mode: None` batches on list of *prompts*. (Batch mode disabled atm, only animation_prompts are working)": "`animation_mode: None` batches en lista de *prompts*. (Modo batch deshabilitado por el momento, solamente animation_prompts esta funcionando)",
"*Important change from vanilla Deforum!*": "*Cambios importantes en Deforum!*",
"This script uses the built-in webui weighting settings.": "Este script utiliza la configuración de pesos integrados.",
"So if you want to use math functions as prompt weights,": "Entonces, si deseas usar funciones matemáticas con pesos en los prompts,",
"keep the values above zero in both parts": "manten los valores por encima de cero en ambas partes",
"Negative prompt part can be specified with --neg": "La parte de prompt negativo se puede especificar utilizando --neg",
"batch_prompts (disabled atm)": "batch_prompts (deshabilitado por el momento)",
"animation_prompts": "animation_prompts",
"Init settings": "Ajustes Init",
"use_init": "use_init",
"from_img2img_instead_of_link": "from_img2img_instead_of_link",
"strength_0_no_init": "strength_0_no_init",
"strength": "strength",
"init_image": "init_image",
"use_mask": "use_mask",
"use_alpha_as_mask": "use_alpha_as_mask",
"invert_mask": "invert_mask",
"overlay_mask": "overlay_mask",
"mask_file": "mask_file",
"mask_brightness_adjust": "mask_brightness_adjust",
"mask_overlay_blur": "mask_overlay_blur",
"Video Input:": "Entrada de Video:",
"video_init_path": "video_init_path",
"extract_nth_frame": "extract_nth_frame",
"overwrite_extracted_frames": "overwrite_extracted_frames",
"use_mask_video": "use_mask_video",
"video_mask_path": "video_mask_path",
"Interpolation (turned off atm)": "Interpolación (apagado por el momento)",
"interpolate_key_frames": "interpolate_key_frames",
"interpolate_x_frames": "interpolate_x_frames",
"Resume animation:": "Reanudar animación:",
"resume_from_timestring": "resume_from_timestring",
"resume_timestring": "resume_timestring",
"Video output settings": "Ajustes video de salida",
"skip_video_for_run_all": "skip_video_for_run_all",
"fps": "fps",
"output_format": "output_format",
"PIL gif": "PIL gif",
"FFMPEG mp4": "FFMPEG mp4",
"ffmpeg_location": "ffmpeg_location",
"add_soundtrack": "add_soundtrack",
"soundtrack_path": "soundtrack_path",
"use_manual_settings": "use_manual_settings",
"render_steps": "render_steps",
"max_video_frames": "max_video_frames",
"path_name_modifier": "path_name_modifier",
"x0_pred": "x0_pred",
"x": "x",
"image_path": "image_path",
"mp4_path": "mp4_path",
"Click here after the generation to show the video": "Haz click aquí después de la generación para mostrar el video",
"NOTE: If the 'Generate' button doesn't work, go in Settings and click 'Restart Gradio and Refresh...'.": "NOTA: Si el botón 'Generar' no funciona, ve a los Ajustes y presiona 'Reinciar Gradio y Refrescar...'.",
"Save Settings": "Guardar Ajustes",
"Load Settings": "Cargar Ajustes",
"Path relative to the webui folder": "Ruta relativa al folder principal",
"Save Video Settings": "Guardar Ajustes de Video",
"Load Video Settings": "Cargar Ajustes de Video",
"Favorites": "Favoritos",
"Others": "Otros",
"Images directory": "Directorio de Imágenes",
"Dropdown": "Menú desplegable",
"First Page": "Primera Página",
"Prev Page": "Página Anterior",
"Page Index": "Índice de Página",
"Next Page": "Página Siguiente",
"End Page": "Última Página",
"delete next": "eliminar siguiente",
"Delete": "Eliminar",
"sort by": "ordenar por",
"path name": "nombre de ruta",
"date": "fecha",
"keyword": "palabra clave",
"Generate Info": "Generar Info",
"File Name": "Nombre de Archivo",
"Move to favorites": "Mover a favoritos",
"Renew Page": "Recargar Página",
"Number": "Número",
"set_index": "set_index",
"load_switch": "load_switch",
"turn_page_switch": "turn_page_switch",
"Checkbox": "Checkbox",
"Apply settings": "Aplicar ajustes",
"Saving images/grids": "Guardar imágenes/grids",
"Always save all generated images": "Siempre guardar imágenes generadas",
"File format for images": "Formato de archivo para imágenes",
"Images filename pattern": "Patrón nombre archivo imágenes",
"Images filename pattern": "Patrón en nombre archivo imágenes",
"Add number to filename when saving": "Agregar número al nombre de archivo al guardar",
"Always save all generated image grids": "Siempre guardar grids de imágenes generadas",
"File format for grids": "Formato de archivo para grids",
@ -262,6 +490,8 @@
"Save text information about generation parameters as chunks to png files": "Guardar información de texto sobre parámetros de generación como fragmentos en archivos png",
"Create a text file next to every image with generation parameters.": "Crear un archivo de texto junto a cada imagen con parámetros de generación.",
"Save a copy of image before doing face restoration.": "Guardar una copia de la imagen antes de restaurar rostro.",
"Save a copy of image before applying highres fix.": "Guardar una copia de la imagen antes de aplicar highres fix.",
"Save a copy of image before applying color correction to img2img results": "Guarda una copia de la imagen antes de aplicar la corrección de color a los resultados de img2img",
"Quality for saved jpeg images": "Calidad para imágenes jpeg guardadas",
"If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG": "Si la imagen PNG es más grande de 4 MB o cualquier dimensión es más grande que 4000, reduce la escala y guarda la copia como JPG",
"Use original name for output filename during batch process in extras tab": "Use el nombre original para el nombre del archivo de salida durante el batch process en la pestaña extras",
@ -291,23 +521,26 @@
"Upscaler for img2img": "Upscaler para img2img",
"Upscale latent space image when doing hires. fix": "Upscale latent space al aplicar hires. fix",
"Face restoration": "Restauración de rostro",
"CodeFormer weight parameter; 0 = maximum effect; 1 = minimum effect": "CodeFormer parámetro peso; 0 = máximo efecto; 1 = mínimo efecto",
"CodeFormer weight parameter; 0 = maximum effect; 1 = minimum effect": "Parámetro influencia CodeFormer; 0 = máximo efecto; 1 = mínimo efecto",
"Move face restoration model from VRAM into RAM after processing": "Mover modelo de restauración de rostro del VRAM al RAM después de procesar",
"System": "Sistema",
"VRAM usage polls per second during generation. Set to 0 to disable.": "Sondeos de uso de VRAM por segundo durante la generación. Establecer en 0 para deshabilitar.",
"Always print all generation info to standard output": "Imprime siempre toda la información de generación en la salida estándar",
"Add a second progress bar to the console that shows progress for an entire job.": "Agrega una segunda barra de progreso a la consola que muestra el progreso de un trabajo completo.",
"Training": "Entrenamiento",
"Move VAE and CLIP to RAM when training hypernetwork. Saves VRAM.": "Mover VAE y CLIP al RAM al entrenar un hypernetwork. Ahorra VRAM.",
"Move VAE and CLIP to RAM when training if possible. Saves VRAM.": "Mover VAE y CLIP al RAM al entrenar cuando sea posible. Ahorra VRAM.",
"Filename word regex": "Filename word regex",
"Filename join string": "Filename join string",
"Number of repeats for a single input image per epoch; used only for displaying epoch number": "Número de repeticiones para una sola imagen de entrada por epoch; utilizado solo para mostrar el número epoch",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "Guarda un csv que contenga la pérdida en el directorio log cada N pasos, 0 para deshabilitar",
"Use cross attention optimizations while training": "Utiliza optimizaciones de atención cruzada durante el entrenamiento",
"Stable Diffusion": "Stable Diffusion",
"Checkpoints to cache in RAM": "Checkpoints al cache en RAM",
"SD VAE": "SD VAE",
"auto": "auto",
"Hypernetwork strength": "Hypernetwork strength",
"Inpainting conditioning mask strength": "Fuerza de la máscara en acondicionamiento Inpainting",
"Apply color correction to img2img results to match original colors.": "Aplica la corrección de color a los resultados de img2img para que coincidan con los colores originales.",
"Save a copy of image before applying color correction to img2img results": "Guarda una copia de la imagen antes de aplicar la corrección de color a los resultados de img2img",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "Con img2img, hace exactamente la cantidad de pasos que especifica el slider (normalmente haría menos con menos eliminación de ruido).",
"Enable quantization in K samplers for sharper and cleaner results. This may change existing seeds. Requires restart to apply.": "Habilita la cuantificación en K samplers para obtener resultados más nítidos y limpios. Esto puede cambiar los seeds existentes. Requiere reiniciar para aplicar.",
"Emphasis: use (text) to make model pay more attention to text and [text] to make it pay less attention": "Énfasis: utiliza (texto) para que el modelo preste más atención al texto y [texto] para que preste menos atención",
@ -324,7 +557,7 @@
"Interrogate: minimum description length (excluding artists, etc..)": "Interrogar: longitud mínima de la descripción (excluyendo artistas, etc.)",
"Interrogate: maximum description length": "Interrogar: longitud máxima de la descripción",
"CLIP: maximum number of lines in text file (0 = No limit)": "CLIP: número máximo de líneas en el archivo de texto (0 = Sin límite)",
"Interrogate: deepbooru score threshold": "Interrogar: deepbooru score threshold",
"Interrogate: deepbooru score threshold": "Interrogar: deepbooru umbral de puntuación",
"Interrogate: deepbooru sort alphabetically": "Interrogar: deepbooru ordenar alfabéticamente",
"use spaces for tags in deepbooru": "usar espacios para etiquetas en deepbooru",
"escape (\\) brackets in deepbooru (so they are used as literal brackets and not for emphasis)": "corchetes de escape (\\) en deepbooru (por lo que se usan como corchetes literales y no para enfatizar)",
@ -337,13 +570,26 @@
"Add model hash to generation information": "Agregar hash de modelo a la información de generación",
"Add model name to generation information": "Agregar nombre de modelo a la información de generación",
"When reading generation parameters from text into UI (from PNG info or pasted text), do not change the selected model/checkpoint.": "Al leer los parámetros de generación del texto en la interfaz de usuario (desde PNG Info o texto pegado), no cambia el modelo/checkpoint seleccionado.",
"Send seed when sending prompt or image to other interface": "Enviar seed cuando se envíe el prompt o imagen a otra interfase",
"Font for image grids that have text": "Tipografía para grids de imágenes que tienen texto",
"Enable full page image viewer": "Habilitar visor de imágenes de página completa",
"Show images zoomed in by default in full page image viewer": "Mostrar imágenes ampliadas de forma predeterminada en el visor de imágenes de página completa",
"Show generation progress in window title.": "Muestra el progreso de la generación en el título de la ventana del navegador.",
"Quicksettings list": "Lista de ajustes rápidos",
"Localization (requires restart)": "Traducción (requiere reiniciar)",
"Sampler parameters": "Parámetros sampler",
"fr_FR": "fr_FR",
"tr_TR": "tr_TR",
"it_IT": "it_IT",
"de_DE": "de_DE",
"ru_RU": "ru_RU",
"ja_JP": "ja_JP",
"es_ES": "es_ES",
"ko_KR": "ko_KR",
"zh_TW": "zh_TW",
"zh_CN": "zh_CN",
"ar_AR": "ar_AR",
"pt_BR": "pt_BR",
"Sampler parameters": "Parámetros del sampler",
"Hide samplers in user interface (requires restart)": "Ocultar samplers en interfaz de usuario (requiere reiniciar)",
"eta (noise multiplier) for DDIM": "eta (noise multiplier) para DDIM",
"eta (noise multiplier) for ancestral samplers": "eta (noise multiplier) para ancestral samplers",
@ -354,10 +600,35 @@
"sigma tmin": "sigma tmin",
"sigma noise": "sigma noise",
"Eta noise seed delta": "Eta noise seed delta",
"Images Browser": "Navegador de Imágenes",
"Preload images at startup": "Precargar imágenes al iniciar",
"Number of columns on the page": "Número de columnas en la página",
"Number of rows on the page": "Número de filas en la página",
"Minimum number of pages per load": "Número mínimo de páginas por carga",
"Request browser notifications": "Solicitar notificaciones del navegador",
"Download localization template": "Descargar plantilla de traducción",
"Reload custom script bodies (No ui updates, No restart)": "Recargar custom script bodies (Sin actualizar UI, Sin reiniciar)",
"Restart Gradio and Refresh components (Custom Scripts, ui.py, js and css only)": "Reiniciar Gradio y Refrescar componentes (Custom Scripts, ui.py, js y css)",
"Installed": "Instaladas",
"Available": "Disponibles",
"Install from URL": "Instalar desde URL",
"Apply and restart UI": "Aplicar y reiniciar UI",
"Check for updates": "Buscar actualizaciones",
"Extension": "Extensión",
"URL": "URL",
"Update": "Actualizar",
"deforum-for-automatic1111-webui": "deforum-for-automatic1111-webui",
"https://github.com/deforum-art/deforum-for-automatic1111-webui": "https://github.com/deforum-art/deforum-for-automatic1111-webui",
"unknown": "desconocido",
"stable-diffusion-webui-images-browser": "stable-diffusion-webui-images-browser",
"https://github.com/yfszzx/stable-diffusion-webui-images-browser": "https://github.com/yfszzx/stable-diffusion-webui-images-browser",
"Load from:": "Cargar desde:",
"Extension index URL": "URL índice de extensiones",
"URL for extension's git repository": "URL repositorio git de extensión",
"Local directory name": "Nombre directorio local",
"Install": "Instalar",
"Ver": "Ver",
"Entrenar un embedding o Hypernetwork; debes especificar un directorio con un conjunto de imágenes con una proporción de 1:1": "Entrenar un embedding o Hypernetwork; debes especificar un directorio con un conjunto de imágenes con una proporción de 1:1",
"Prompt (press Ctrl+Enter or Alt+Enter to generate)": "Prompt (presiona Ctrl+Enter o Alt+Enter para generar)",
"Negative prompt (press Ctrl+Enter or Alt+Enter to generate)": "Prompt negativo (presiona Ctrl+Enter o Alt+Enter para generar)",
"Add a random artist to the prompt.": "Agregar un artista aleatorio al prompt.",
@ -387,16 +658,16 @@
"Separate values for Y axis using commas.": "Separar valores para Y usando comas.",
"Write image to a directory (default - log/images) and generation parameters into csv file.": "Escribe la imagen en un directorio (predeterminado: log/images) y los parámetros de generación en el archivo csv.",
"Open images output directory": "Abrir directorio de imágenes",
"How much to blur the mask before processing, in pixels.": "Cuánto blur al mask antes de procesar, en píxeles.",
"What to put inside the masked area before processing it with Stable Diffusion.": "Qué poner dentro del área con mask antes de procesarlo con Stable Diffusion.",
"How much to blur the mask before processing, in pixels.": "Cuánto difuminado a la máscara antes de procesarla, en píxeles.",
"What to put inside the masked area before processing it with Stable Diffusion.": "Qué poner dentro del área enmascarada antes de procesarla con Stable Diffusion.",
"fill it with colors of the image": "rellenarlo con los colores de la imagen",
"keep whatever was there originally": "mantener lo que estaba allí originalmente",
"fill it with latent space noise": "rellenarlo con latent space noise",
"fill it with latent space zeroes": "rellenarlo con latent space zeroes",
"Upscale masked region to target resolution, do inpainting, downscale back and paste into original image": "Upscale la región con mask a la resolución objetivo, vuelve a pintar, reduce la escala hacia atrás y pégala en la imagen original",
"Resize image to target resolution. Unless height and width match, you will get incorrect aspect ratio.": "Cambia el tamaño de la imagen a la resolución destino. A menos que la altura y el ancho coincidan, obtendrás una relación de aspecto incorrecta.",
"Upscale masked region to target resolution, do inpainting, downscale back and paste into original image": "Escalar la región con máscara a la resolución objetivo, vuelve a pintar, reduce la escala hacia atrás y pégala en la imagen original",
"Resize image to target resolution. Unless height and width match, you will get incorrect aspect ratio.": "Cambia el tamaño de la imagen a la resolución objetivo. A menos que la altura y el ancho coincidan, obtendrás una relación de aspecto incorrecta.",
"Resize the image so that entirety of target resolution is filled with the image. Crop parts that stick out.": "Cambia el tamaño de la imagen para que la totalidad de la resolución destino se llene con la imagen. Recorta las partes que sobresalen.",
"Resize the image so that entirety of image is inside target resolution. Fill empty space with image's colors.": "Cambia el tamaño de la imagen para que la totalidad de la imagen esté dentro de la resolución de destino. Rellena el espacio vacío con los colores de la imagen.",
"Resize the image so that entirety of image is inside target resolution. Fill empty space with image's colors.": "Cambia el tamaño de la imagen para que la totalidad de la imagen esté dentro de la resolución objetivo. Rellena el espacio vacío con los colores de la imagen.",
"How many times to repeat processing an image and using it as input for the next iteration": "Cuántas veces repetir el procesamiento de una imagen y usarla como entrada para la próxima iteración",
"In loopback mode, on each loop the denoising strength is multiplied by this value. <1 means decreasing variety so your sequence will converge on a fixed picture. >1 means increasing variety so your sequence will become more and more chaotic.": "En modo loopback, en cada bucle, la fuerza de eliminación de ruido se multiplica por este valor. <1 significa variedad decreciente, por lo que su secuencia convergerá en una imagen fija. >1 significa aumentar la variedad, por lo que su secuencia se volverá cada vez más caótica.",
"For SD upscale, how much overlap in pixels should there be between tiles. Tiles overlap so that when they are merged back into one picture, there is no clearly visible seam.": "Para SD upscale, cuánta superposición en píxeles debe haber entre mosaicos. Los mosaicos se superponen de modo que cuando se fusionan nuevamente en una imagen, no hay una unión claramente visible.",
@ -407,12 +678,15 @@
"1st and last digit must be 1. ex:'1, 2, 1'": "Primer y último dígito debe ser 1. ej:'1, 2, 1'",
"Path to directory with input images": "Ruta al directorio con imágenes de entrada",
"Path to directory where to write outputs": "Ruta al directorio donde escribir salidas",
"Input images directory": "Directorio de imágenes de entrada",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "Usa las siguientes etiquetas para definir cómo se eligen los nombres de archivo para las imágenes: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; dejar vacío para utilizar predeterminados.",
"If this option is enabled, watermark will not be added to created images. Warning: if you do not add watermark, you may be behaving in an unethical manner.": "Si esta opción está habilitada, el watermark no se agregará a las imágenes creadas. Advertencia: si no agregas un watermark, es posible que te estés comportando de manera poco ética.",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "Usa las siguiente etiquetas para definir cómo los subdirectorios para imágenes y grids son seleccionados: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; dejar vacío para utilizar predeterminados.",
"Restore low quality faces using GFPGAN neural network": "Restaurar rostros de baja calidad utilizando GFPGAN neural network",
"This regular expression will be used extract words from filename, and they will be joined using the option below into label text used for training. Leave empty to keep filename text as it is.": "Esta expresión regular se usará para extraer palabras del nombre de archivo y se unirán usando la opción a continuación en el texto de la etiqueta que se usa para el entrenamiento. Dejar vacío para mantener el texto del nombre de archivo tal como está.",
"This string will be used to join split words into a single line if the option above is enabled.": "Esta cadena se usará para unir palabras divididas en una sola línea si la opción anterior está habilitada.",
"Only applies to inpainting models. Determines how strongly to mask off the original image for inpainting and img2img. 1.0 means fully masked, which is the default behaviour. 0.0 means a fully unmasked conditioning. Lower values will help preserve the overall composition of the image, but will struggle with large changes.": "Solo se aplica a modelos inpainting. Determina con qué fuerza enmascarar la imagen original para inpainting en img2img. 1.0 significa totalmente enmascarado, que es el comportamiento predeterminado. 0.0 significa un condicionamiento totalmente desenmascarado. Los valores más bajos ayudarán a preservar la composición general de la imagen, pero tendrán problemas con los grandes cambios.",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "Lista de nombres de configuración, separados por comas, para configuraciones que deben ir a la barra de acceso rápido en la parte superior, en lugar de la pestaña de configuración habitual. Ver modules/shared.py para configurar los nombres. Requiere reiniciar para aplicar.",
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "Si este valor no es cero, se agregará al seed y se usará para inicializar RNG para ruidos cuando se usan samplers con Eta. Puedes usar esto para producir aún más variaciones de imágenes, o puedes usar esto para hacer coincidir imágenes de otro software si sabes lo que estás haciendo."
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "Si este valor no es cero, se agregará al seed y se usará para inicializar RNG para ruidos cuando se usan samplers con Eta. Puedes usar esto para producir aún más variaciones de imágenes, o puedes usar esto para hacer coincidir imágenes de otro software si sabes lo que estás haciendo.",
"Leave empty for auto": "Dejar vacío para automático"
}

View file

@ -17,11 +17,14 @@
"Checkpoint Merger": "Miscelatore di Checkpoint",
"Train": "Addestramento",
"Create aesthetic embedding": "Crea incorporamento estetico",
"auto-sd-paint-ext Guide/Panel": "KRITA Plugin Guida",
"Dataset Tag Editor": "Dataset Tag Editor",
"Deforum": "Deforum",
"Artists To Study": "Artisti per studiare",
"Image Browser": "Galleria immagini",
"Inspiration": "Ispirazione",
"Tokenizer": "Tokenizzatore",
"Training Picker": "Training Picker",
"Settings": "Impostazioni",
"Extensions": "Estensioni",
"Prompt": "Prompt",
@ -49,11 +52,15 @@
"Heun": "Heun",
"DPM2": "DPM2",
"DPM2 a": "DPM2 a",
"DPM++ 2S a": "DPM++ 2S a",
"DPM++ 2M": "DPM++ 2M",
"DPM fast": "DPM fast",
"DPM adaptive": "DPM adaptive",
"LMS Karras": "LMS Karras",
"DPM2 Karras": "DPM2 Karras",
"DPM2 a Karras": "DPM2 a Karras",
"DPM++ 2S a Karras": "DPM++ 2S a Karras",
"DPM++ 2M Karras": "DPM++ 2M Karras",
"DDIM": "DDIM",
"PLMS": "PLMS",
"Width": "Larghezza",
@ -84,19 +91,30 @@
"Aesthetic text for imgs": "Estetica - Testo per le immagini",
"Slerp angle": "Angolo Slerp",
"Is negative text": "È un testo negativo",
"Unprompted": "Unprompted",
"Dry Run": "Esecuzione a vuoto (Debug)",
"NEW!": "NUOVO!",
"Premium Fantasy Card Template": "Premium Fantasy Card Template",
"is now available.": "è ora disponibile.",
"Generate a wide variety of creatures and characters in the style of a fantasy card game. Perfect for heroes, animals, monsters, and even crazy hybrids.": "Genera un'ampia varietà di creature e personaggi nello stile di un gioco di carte fantasy. Perfetto per eroi, animali, mostri e persino ibridi incredibili.",
"Learn More ➜": "Per saperne di più ➜",
"Purchases help fund the continued development of Unprompted. Thank you for your support!": "Gli acquisti aiutano a finanziare il continuo sviluppo di Unprompted. Grazie per il vostro sostegno!",
"Dismiss": "Rimuovi",
"Script": "Script",
"Random grid": "Generaz. casuale (griglia)",
"Random": "Generaz. casuale (no griglia)",
"StylePile": "StylePile",
"Advanced prompt matrix": "Matrice di prompt avanzata",
"Advanced Seed Blending": "Miscelazione Semi Avanzata",
"Alternate Sampler Noise Schedules": "Metodi alternativi di campionamento del rumore",
"Animator v6": "Animator v6",
"Asymmetric tiling": "Piastrellatura asimmetrica",
"SD Chad - Stable Diffusion Aesthetic Scorer": "SD Chad - Assegna Punteggio Estetica",
"Custom code": "Codice personalizzato",
"Dynamic Prompting v0.2": "Prompt dinamici v0.2",
"Embedding to Shareable PNG": "Incorporamento convertito in PNG condivisibile",
"DepthMap v0.1.0": "DepthMap v0.1.0",
"Embedding to Shareable PNG": "Converte Incorporamento in PNG condivisibile",
"Force symmetry": "Forza la simmetria",
"Improved prompt matrix": "Improved prompt matrix",
"Latent Mirroring": "Rispecchia e capovolge l'immagine latente",
"Prompts interpolation": "Interpola Prompt",
"Prompt matrix": "Matrice dei prompt",
"Prompt morph": "Metamorfosi del prompt",
@ -107,6 +125,7 @@
"Text to Vector Graphics": "Da testo a grafica vettoriale",
"X/Y plot": "Grafico X/Y",
"X/Y/Z plot": "Grafico X/Y/Z",
"Dynamic Prompting v0.13.6": "Prompt dinamici v0.13.6",
"Create inspiration images": "Crea immagini di ispirazione",
"step1 min/max": "Passi min(o max)",
"step2 min/max": "Passi max (o min)",
@ -119,70 +138,186 @@
"Keep -1 for seeds": "Mantieni sempre il seme a -1",
"x/y change": "Inverti ordine assi X/Y (Passi/CFG)",
"Loops": "Cicli",
"Focus on:": "Focus su:",
"No focus": "Nessun Focus",
"Portraits (tick Restore faces above for best results)": "Ritratti (selezionare 'Restaura volti' in alto per ottenere i migliori risultati)",
"Feminine and extra attractive (tick Restore faces above for best results)": "Femminile ed estremamente attraente (selezionare 'Restaura volti' per ottenere i migliori risultati)",
"Masculine and extra attractive (tick Restore faces above for best results)": "Maschile ed estremamente attraente (selezionare 'Restaura volti' per ottenere i migliori risultati)",
"Monsters": "Mostri",
"Robots": "Robot",
"Retrofuturistic": "Retrofuturistico",
"Propaganda": "Propaganda",
"Landscapes": "Paesaggi",
"Hints": "Suggerimenti",
"Set Batch count to Sequential prompt count": "Imposta il lotto di immagini sul numero di prompt sequenziali",
"Sequential prompts [X]": "Prompt sequenziali [X]",
"Random prompts [R]": "Prompt randomici [R]",
"Image type strength": "Intensità del tipo di immagine",
"Image type": "Tipo di immagine",
"Not set": "Non impostato",
"Photography": "Fotografia",
"Digital art": "Arte digitale",
"3D Rendering": "3D Rendering",
"3D Rendering": "Rendering 3D",
"Painting": "Dipinto",
"Sketch": "Schizzo",
"Classic Comics": "Fumetti classici",
"Modern Comics": "Fumetti moderni",
"Manga": "Manga",
"Vector art": "Arte vettoriale",
"Visual style": "Stile visivo",
"Realism": "Realismo",
"Photorealism": "Fotorealismo",
"Hyperrealism": "Iperrealismo",
"Surrealism": "Surrealismo",
"Modern Art": "Arte moderna",
"Fauvism": "Fauvismo",
"Futurism": "Futurismo",
"Painterly": "Pittorico",
"Pointillisme": "Puntinismo",
"Abstract": "Astratto",
"Pop Art": "Pop Art",
"Impressionist": "Impressionista",
"Cubism": "Cubismo",
"Linocut": "Linoleografia",
"Fantasy": "Fantasia",
"Vector Art": "Arte vettoriale",
"Mood": "Stato d'animo",
"Amusing": "Amusing",
"Angry": "Angry",
"Cosy": "Cosy",
"Depressing": "Depressing",
"Disgusting": "Disgusting",
"Embarrassing": "Embarrassing",
"Energetic": "Energetic",
"Evil": "Evil",
"Fearful": "Fearful",
"Frightening": "Frightening",
"Grim": "Grim",
"Guilty": "Guilty",
"Happy": "Happy",
"Hopeful": "Hopeful",
"Hopeless": "Hopeless",
"Lonely": "Lonely",
"Lustful": "Lustful",
"Peaceful": "Peaceful",
"Proud": "Proud",
"Relieving": "Relieving",
"Romantic": "Romantic",
"Sad": "Sad",
"Satisfying": "Satisfying",
"Shameful": "Shameful",
"Surprising": "Surprising",
"Colors": "Colori",
"Chaotic": "Caotico",
"Primary colors": "Colori primari",
"Colorful": "Colorato",
"Primary Colors": "Primary Colors",
"Vivid": "Vivido",
"Muted colors": "Colori tenui",
"Low contrast": "Basso contrasto",
"Desaturated": "Desaturato",
"Grayscale": "Scala di grigi",
"Black and white": "Bianco e nero",
"Infrared": "Infrarosso",
"Complementary": "Colori complementari",
"Non-complementary": "Colori non complementari",
"Pastel Colors": "Pastel Colors",
"Muted Colors": "Muted Colors",
"Grayscale": "Grayscale",
"Black and white": "Black and white",
"Infrared": "Infrared",
"View": "Visuale",
"Tilt shift": "Tilt shift",
"Wide-angle": "Angolo ampio",
"Symmetrical": "Symmetrical",
"Tilt-shift": "Tilt-shift",
"Long shot angle": "Long shot angle",
"Medium shot angle": "Medium shot angle",
"Wide shot angle": "Wide shot angle",
"Portrait": "Ritratto",
"Extreme close-up angle": "Extreme close-up angle",
"Macro": "Macro",
"Microscopic": "Microscopico",
"Isometric": "Isometrico",
"Panorama": "Panorama",
"Aerial photograph": "Fotografia aerea",
"Artist focus (not quite finished, not sure it helps)": "Focus sull'artista (non del tutto finito, non è sicuro che aiuti)",
"B/W Photograpy": "Fotografia B/N",
"Portrait photo": "Foto ritratto",
"Usage: a <corgi|cat> wearing <goggles|a hat>": "Utilizzo: a <corgi|cat> wearing <goggles|a hat>",
"Fisheye lens": "Fisheye lens",
"Overhead-angle": "Overhead-angle",
"Birds eye view": "Birds eye view",
"Focus on (unfinished)": "Focus su (incompleto)",
"No focus": "Nessun focus",
"Portraits": "Portraits",
"Feminine portrait": "Ritratto femminile",
"Masculine portrait": "Ritratto maschile",
"WaiFusion": "WaiFusion",
"Horrible Monsters": "Horrible Monsters",
"Robots": "Robot",
"Retrofuturism": "Retrofuturism",
"Propaganda": "Propaganda",
"Landscapes": "Paesaggi",
"Visual style strength": "Forza dello stile visivo",
"Visual style": "Stile visivo",
"Abstract": "Abstract",
"Acidwave": "Acidwave",
"Acrylic": "Acrylic",
"Aestheticism": "Aestheticism",
"Art Deco": "Art Deco",
"Nouveau": "Nouveau",
"Ashcan School Style": "Ashcan School Style",
"Avant-garde": "Avant-garde",
"Ballpoint Pen": "Ballpoint Pen",
"Baroque": "Baroque",
"Classicism": "Classicism",
"CoBrA": "CoBrA",
"Colored Pencil": "Colored Pencil",
"Constructivism": "Constructivism",
"Cubism": "Cubism",
"Dreamcore": "Dreamcore",
"Drip": "Drip",
"Encaustic": "Encaustic",
"Expressionism": "Expressionism",
"Fauvism": "Fauvism",
"Finger Painting": "Finger Painting",
"Futurism": "Futurism",
"Gouache": "Gouache",
"Hot Wax": "Hot Wax",
"Impressionism": "Impressionism",
"Ink Wash": "Ink Wash",
"Japanese": "Japanese",
"Korean": "Korean",
"Line Art": "Line Art",
"Linocut": "Linocut",
"Lowpoly": "Lowpoly",
"Manga": "Manga",
"Marker": "Marker",
"Modern Comics": "Fumetti moderni",
"Mural": "Mural",
"Neoclassicism": "Neoclassicism",
"Oil Painting": "Oil Painting",
"Pastel": "Pastel",
"Pencil": "Pencil",
"Photorealism": "Photorealism",
"Pixel Art": "Pixel Art",
"Pop Art": "Pop Art",
"Pop Surrealism": "Pop Surrealism",
"Psychedelic": "Psychedelic",
"Realism": "Realism",
"Renaissance": "Renaissance",
"Rococo": "Rococo",
"Sprite Artwork": "Sprite Artwork",
"Street Art": "Street Art",
"Suprematism": "Suprematism",
"Vaporwave": "Vaporwave",
"Vintage Comics": "Vintage Comics",
"Watercolor": "Watercolor",
"Artist influence": "Influenza dell'Artista",
"Artist": "Artista",
"Examples, Help and Tips": "Esempi, aiuto e suggerimenti",
"Style examples": "Esempi di Stile",
"Artist examples": "Esempi di Artisti",
"Mood examples": "Esempi di stato d'animo",
"Help": "Aiuto",
"Hello, StylePile here": "Ciao, questo è StylePile",
"Introduction": "Introduzione",
"is a mix and match system for adding elements to prompts that affect the style of the result. Hence the name. By default, these elements are placed in a specific order and given strength values. Which means the result sort-of evolves. I have generated thousands of images for each main": "è un sistema mescola e abbina per aggiungere elementi ai prompt che influiscono sullo stile del risultato. Da qui il nome. Per impostazione predefinita, questi elementi vengono inseriti in un ordine specifico e vengono assegnati valori di forza. Il che significa che il risultato si evolve. Ho generato migliaia di immagini per ogni principale",
"and tweaked the keywords to attempt giving expected results most of the time. Certainly, your suggestions for improvements are very welcome.": "e ottimizzato le parole chiave per tentare di fornire i risultati attesi per la maggior parte del tempo. Certamente, i tuoi suggerimenti per miglioramenti sono molto graditi.",
"Base workflow": "Flusso di lavoro di base",
"For example, if you select the": "Ad esempio, se si seleziona il",
"image type, then almost all results will look like Paintings. Selecting": "tipo di immagine, quindi quasi tutti i risultati sembreranno Dipinti. Selezione",
"will have a certain influence on the overall look in some way (if it's something humanoid it may show emotion, but also colors and overall feel may change). Setting": "avrà una certa influenza sull'aspetto generale in qualche modo (se è qualcosa di umanoide può mostrare emozioni, ma anche i colori e la sensazione generale possono cambiare). Ambientazione",
"will change the general tonality of the result. And setting": "cambierà la tonalità generale del risultato. E ambientazione",
"will attempt to change how the subject is viewed. Attempt, because view appears to be the least reliable keyword. These elements are placed in order of influence and supported by certain strength values. These basic settings produce very quick results close to the general look you want.": "tenterà di cambiare la modalità di visualizzazione dell'oggetto. Tentativo, perché view sembra essere la parola chiave meno affidabile. Questi elementi sono posti in ordine di influenza e supportati da determinati valori di forza. Queste impostazioni di base producono risultati molto rapidi vicini all'aspetto generale desiderato.",
"Moving on, adding a": "Andando avanti, aggiungendo a",
"will combine with": "si combinerà con",
"to influence how the result generally looks. These styles are based on classic and modern Painting/Art/design movements (which I picked after hours and thousands of samples of testing) and can have a strong influence on the end result. Either it will be more realistic or artistic, or look like a comic book etc. In general, this is a really strong element for getting the look you want. Its influence can be adjusted with the slider above. Experiment with the values, keeping in mind that anything above 1.5 will start becoming a mess. In a similar way, but more focused, you can select an": "per influenzare l'aspetto generale del risultato. Questi stili si basano su movimenti Pittura/Arte/design classici e moderni (che ho scelto dopo ore e migliaia di campioni di test) e possono avere una forte influenza sul risultato finale. O sarà più realistico o artistico, o assomiglierà a un fumetto, ecc. In generale, questo è un elemento davvero forte per ottenere l'aspetto che desideri. La sua influenza può essere regolata con il cursore in alto. Sperimenta con i valori, tenendo presente che qualsiasi cosa al di sopra di 1,5 inizierà a diventare un pasticcio. In modo simile, ma più mirato, puoi selezionare un",
"and, of course, that will have a very visible effect on the result as well. Currently there are 135 artists, 55 art styles and 25 emotions available for selection and represented with preview images.": "e, naturalmente, anche questo avrà un effetto molto visibile sul risultato. Attualmente ci sono 135 artisti, 55 stili artistici e 25 emozioni disponibili per la selezione e rappresentate con immagini di anteprima.",
"Strength of these settings has been preset at 1.3, as that appears to be the golden ratio for getting good results. Sometimes very low settings have an interesting result as well. You can, and should, freely mix and match these settings to get different results. Classic Painting styles affected or affecting 3D look quite interesting. Photography can look cool with some of the brighter, more artistic styles etc. Sometimes raising CFG scale to 15,20 or more also helps to REALLY push the style onto the image.": "La forza di queste impostazioni è stata preimpostata a 1,3, poiché sembra essere il rapporto aureo per ottenere buoni risultati. A volte anche impostazioni molto basse hanno un risultato interessante. Puoi, e dovresti, mescolare e abbinare liberamente queste impostazioni per ottenere risultati diversi. Gli stili di pittura classica influenzati o influenzati dal 3D sembrano piuttosto interessanti. La fotografia può sembrare interessante con alcuni degli stili più luminosi, più artistici, ecc. A volte aumentare la scala CFG a 15,20 o più aiuta anche a spingere VERAMENTE lo stile sull'immagine.",
"Advanced workflow": "Flusso di lavoro avanzato",
"StylePile can overtake the generation process, allowing you to generate a large amount of different results with very little extra work. There are two types of variables you can use: [X] and [R]. When you add an [X] to your prompt, it sequentially takes values from the": "StylePile può prendere il controllo del processo di generazione, consentendoti di generare una grande quantità di risultati diversi con pochissimo lavoro extra. Ci sono due tipi di variabili che puoi usare: [X] e [R]. Quando aggiungi una [X] al tuo prompt, prende in sequenza i valori da",
"Sequential prompts": "Prompt sequenziali",
"text area. You can have dozens of lines there and they will be processed in sequence. When you add [R] to the prompt a value from the": "area di testo. Puoi avere dozzine di righe lì e verranno elaborate in sequenza. Quando si aggiunge [R] al prompt un valore da",
"Random prompts": "Prompt casuali",
"text area will be inserted in its place. By combining these a huge variety in prompts is very easy to do.": "l'area di testo verrà inserita al suo posto. Combinando questi un'enorme varietà di prompt è molto facile da fare.",
"When using this,": "Quando si utilizza questo,",
"will move through the prompts and": "si sposterà attraverso i prompt e",
"will set how many copies with the given prompt to make. If the seed is not random, it will increase with each batch size step. Any random elements will still be picked randomly.": "imposterà quante copie fare con il prompt dato. Se il seme non è casuale, aumenterà ad ogni passaggio della dimensione del batch. Eventuali elementi casuali verranno comunque selezionati casualmente.",
"Tips and tricks": "Suggerimenti e trucchi",
"If you add your own Artist, I would recommend having \"by Artist\" in front of their name. Depending on Artist's popularity (or lack thereof) this appears to have a very tangible influence on the result.": "Se aggiungi il tuo artista, ti consiglio di avere \"per artista\" davanti al suo nome. A seconda della popolarità dell'artista (o della sua mancanza) questo sembra avere un'influenza molto tangibile sul risultato.",
"Parenthesis can be added to make pArts of the prompt stronger. So": "È possibile aggiungere parentesi per rendere più forti parti del prompt. Così",
"((cute kitten))": "((cute kitten))",
"will make it extra cute (try it out). This is also important if a style is affecting your original prompt too much. Make that prompt stronger by adding parenthesis around it, like this:": "lo renderà molto carino (provalo). Questo è importante anche se uno stile influisce troppo sul prompt originale. Rendi più forte quel prompt aggiungendo parentesi attorno ad esso, in questo modo:",
"((promt))": "((promt))",
". A strength modifier value can also be used, like this": ". È anche possibile utilizzare un valore modificatore di forza, in questo modo",
"(prompt:1.1)": "(prompt:1.1)",
". To save some typing you can select the line you want to make stronger and use": ". Per salvare un po' di digitazione puoi selezionare la linea che vuoi rendere più forte e utilizzare",
"Ctrl+Shift+Arrow keys up": "Ctrl+Shift+Freccia su",
"to add these parenthesis and change the value. As you can see by default values on most sliders, 1.3 seems like a good stArting point if you want to see some impact.": "per aggiungere queste parentesi e modificare il valore. Come puoi vedere per i valori predefiniti sulla maggior parte dei cursori, 1.3 sembra un buon punto di partenza se vuoi vedere un certo impatto.",
"Prompts can be split like": "I prompt possono essere divisi come",
"[A|B]": "[A|B]",
"to sequentially use terms, one after another on each step. For example": "per utilizzare in sequenza i termini, uno dopo l'altro in ogni passaggio. Per esempio",
"[cat|dog]": "[cat|dog]",
"will produce a hybrid catdog.": "produrrà un cane-gatto ibrido.",
"Using": "Usando",
"[A:B:0.4]": "[A:B:0.4]",
"will switch to other terms after the first one has been active for a certain percentage of steps. So": "passerà ad altri termini dopo che il primo è stato attivo per una certa percentuale di passaggi. Così",
"[cat:dog:0.4]": "[cat:dog:0.4]",
"will build a cat 40% of the time and then stArt turning it into a dog. Usually this needs more steps to work properly.": "costruirà un gatto il 40% delle volte e poi inizierà a trasformarlo in un cane. Di solito questo richiede più passaggi per funzionare correttamente.",
"In conclusion": "In conclusione",
"I made this because manually changing keywords, looking up possible styles, etc was a pain. It is meant as a fun tool to explore possibilities and make learning Stable Diffusion easier. If you have some ideas or, better yet, would like to contribute in some way*, just visit https://github.com/some9000/StylePile\n*Hey, if you have a 12Gb graphics card just laying around I'm happy to take it (:": "L'ho fatto perché cambiare manualmente le parole chiave, cercare possibili stili, ecc. era una seccatura. È inteso come uno strumento divertente per esplorare le possibilità e rendere più facile l'apprendimento della Diffusione stabile. Se hai qualche idea o, meglio ancora, vorresti contribuire in qualche modo*, visita pure https://github.com/some9000/StylePile\n*Hey, se hai una scheda grafica da 12 Gb in giro, sono felice di prenderla (:",
"Seeds": "Semi",
"Noise Scheduler": "Pianificazione del rumore",
"Default": "Predefinito",
@ -195,7 +330,7 @@
"Beta distribution (VP only)": "Distribuzione Beta (Solo CV)",
"Beta min (VP only)": "Beta min (Solo CV)",
"Epsilon (VP only)": "Epsilon (Solo CV)",
"Running in txt2img mode:": "Running in txt2img mode:",
"Running in txt2img mode:": "Esecuzione in modalità txt2img:",
"Render these video formats:": "Renderizza in questi formati:",
"GIF": "GIF",
"MP4": "MP4",
@ -203,11 +338,11 @@
"Animation Parameters": "Parametri animazione",
"Total Animation Length (s)": "Durata totale dell'animazione (s)",
"Framerate": "Frequenza dei fotogrammi",
"Smoothing_Frames": "Smoothing_Frames",
"Smoothing_Frames": "Fotogrammi da appianare",
"Add_Noise": "Aggiungi rumore",
"Noise Strength": "Intensità del rumore",
"Initial Parameters": "Parametri iniziali",
"Denoising Strength (overrides img2img slider)": "Intensità di riduzione del rumore (sovrascrive il cursore img2img)",
"Denoising Strength": "Intensità di riduzione del rumore",
"Seed_March": "Seed_March",
"Zoom Factor (scale/s)": "Fattore di ingrandimento (scala/s)",
"X Pixel Shift (pixels/s)": "Sposta i Pixel sull'asse X (pixel/s)",
@ -216,8 +351,8 @@
"Prompt Template, applied to each keyframe below": "Modello di prompt, applicato a ciascun fotogramma chiave qui di seguito",
"Positive Prompts": "Prompt positivi",
"Negative Prompts": "Prompt negativi",
"Props, Stamps": "Props, Stamps",
"Poper_Folder:": "Cartella Poper:",
"Props, Stamps": "Immagini Clipart da diffondere (prop), o da applicare in post elaborazione e non essere diffuse (stamp)",
"Poper_Folder:": "Cartella Immagini Clipart (PNG trasparenti):",
"Supported Keyframes:": "Fotogrammi chiave supportati:",
"time_s | source | video, images, img2img | path": "time_s | source | video, images, img2img | path",
"time_s | prompt | positive_prompts | negative_prompts": "time_s | prompt | positive_prompts | negative_prompts",
@ -234,26 +369,26 @@
"time_s | col_set": "time_s | col_set",
"time_s | col_clear": "time_s | col_clear",
"time_s | model | model": "time_s | model | model",
"img2img_mode": "img2img_mode",
"img2img_mode": "Modalità img2img",
"Keyframes:": "Fotogrammi chiave:",
"Tile X": "Piastrella asse X",
"Tile Y": "Piastrella asse Y",
"Number of images to generate": "Numero di immagini da generare",
"Python code": "Codice Python",
"Combinatorial generation": "Generazione combinatoria",
"Combinations": "Combinazioni",
"Choose a number of terms from a list, in this case we choose two artists": "Scegli un numero di termini da un elenco, in questo caso scegliamo due artisti",
"{2$$artist1|artist2|artist3}": "{2$$artist1|artist2|artist3}",
"If $$ is not provided, then 1$$ is assumed.\n\n A range can be provided:": "Se $$ non viene fornito, si presume 1$$.\n\n È possibile fornire un intervallo di valori:",
"{1-3$$artist1|artist2|artist3}": "{1-3$$artist1|artist2|artist3}",
"In this case, a random number of artists between 1 and 3 is chosen.": "In questo caso viene scelto un numero casuale di artisti compreso tra 1 e 3.",
"Wildcards": "Termini jolly",
"If the groups wont drop down click": "Se i gruppi non vengono visualizzati, clicca",
"here": "qui",
"to fix the issue.": "per correggere il problema.",
"WILDCARD_DIR: scripts/wildcards": "WILDCARD_DIR: scripts/wildcards",
"You can add more wildcards by creating a text file with one term per line and name is mywildcards.txt. Place it in scripts/wildcards.": "Puoi aggiungere termini jolly creando un file di testo con un termine per riga e nominandolo, per esempio, mywildcards.txt. Inseriscilo in scripts/wildcards.",
"__<folder>/mywildcards__": "__<folder>/mywildcards__",
"will then become available.": "diverrà quindi disponibile.",
"Model": "Modello",
"dpt_large": "Depth large",
"dpt_hybrid": "Depth hybrid",
"midas_v21": "MIDAS v21",
"midas_v21_small": "MIDAS v21 small",
"Compute on": "Calcola",
"GPU": "GPU",
"CPU": "CPU",
"Save DepthMap": "Salva DepthMap",
"Show DepthMap": "Mostra DepthMap",
"Combine into one image.": "Combina in un'unica immagine.",
"Combine axis": "Combina gli assi",
"Vertical": "Verticale",
"Horizontal": "Orizzontale",
"Source embedding to convert": "Incorporamento sorgente da convertire",
"Embedding token": "Token Incorporamento",
"Output directory": "Cartella di output",
@ -262,6 +397,16 @@
"Alt. symmetry method (blending)": "Metodo di simmetria alternativo (miscelazione)",
"Apply every n steps": "Applica ogni n passi",
"Skip last n steps": "Salta gli ultimi n passi",
"Usage: a <corgi|cat> wearing <goggles|a hat>": "Utilizzo: a <corgi|cat> wearing <goggles|a hat>",
"Mirror application mode": "modalità di applicazione specchiatura",
"Alternate Steps": "Passaggi alternati",
"Blend Average": "Miscelazione media",
"Mirror style": "Stile specchiatura",
"Vertical Mirroring": "Specchiatura verticale",
"Horizontal Mirroring": "Specchiatura orizzontale",
"90 Degree Rotation": "Rotazione 90 gradi",
"180 Degree Rotation": "Rotazione 180 gradi",
"Maximum steps fraction to mirror at": "Frazione di passi massima su cui eseguire la specchiatura",
"Interpolation prompt": "Prompt di interpolazione",
"Number of images": "Numero di immagini",
"Make a gif": "Crea GIF",
@ -274,9 +419,10 @@
"Save results as video": "Salva i risultati come video",
"Frames per second": "Fotogrammi al secondo",
"Iterate seed every line": "Iterare il seme per ogni riga",
"Use same random seed for all lines": "Usa lo stesso seme casuale per tutte le righe",
"List of prompt inputs": "Elenco di prompt di input",
"Upload prompt inputs": "Carica un file contenente i prompt di input",
"n": "Esegui n volte",
"n": "n",
"Destination seed(s) (Comma separated)": "Seme/i di destinazione (separati da virgola)",
"Only use Random seeds (Unless comparing paths)": "Usa solo semi casuali (a meno che non si confrontino i percorsi)",
"Number of random seed(s)": "Numero di semi casuali",
@ -324,19 +470,37 @@
"Eta": "ETA",
"Clip skip": "Salta CLIP",
"Denoising": "Riduzione del rumore",
"Cond. Image Mask Weight": "Cond. Image Mask Weight",
"Cond. Image Mask Weight": "Peso maschera immagine condizionale",
"X values": "Valori per X",
"Y type": "Parametro asse Y",
"Y values": "Valori per Y",
"Z type": "Parametro asse Z",
"Z values": "Valori per Z",
"Combinatorial generation": "Generazione combinatoria",
"Combinatorial batches": "Lotti combinatori",
"Magic prompt": "Prompt magico",
"Fixed seed": "Seme fisso",
"Combinations": "Combinazioni",
"Choose a number of terms from a list, in this case we choose two artists": "Scegli un numero di termini da un elenco, in questo caso scegliamo due artisti",
"{{2$artist1|artist2|artist3}}": "{{2$artist1|artist2|artist3}}",
"If $ is not provided, then 1$ is assumed.\n\n A range can be provided:": "Se $ non viene fornito, si presume 1$.\n\n È possibile fornire un intervallo di valori:",
"{{1-3$artist1|artist2|artist3}}": "{{1-3$artist1|artist2|artist3}}",
"In this case, a random number of artists between 1 and 3 is chosen.": "In questo caso viene scelto un numero casuale di artisti compreso tra 1 e 3.",
"Wildcards": "Termini jolly",
"If the groups wont drop down click": "Se i gruppi non vengono visualizzati, clicca",
"here": "qui",
"to fix the issue.": "per correggere il problema.",
"WILDCARD_DIR: C:\\stable-diffusion-webui\\extensions\\sd-dynamic-prompts\\wildcards": "WILDCARD_DIR: C:\\stable-diffusion-webui\\extensions\\sd-dynamic-prompts\\wildcards",
"You can add more wildcards by creating a text file with one term per line and name is mywildcards.txt. Place it in scripts/wildcards.": "Puoi aggiungere termini jolly creando un file di testo con un termine per riga e nominandolo, per esempio, mywildcards.txt. Inseriscilo in scripts/wildcards.",
"__<folder>/mywildcards__": "__<cartella>/mywildcards__",
"will then become available.": "diverrà quindi disponibile.",
"Artist or styles name list. '.txt' files with one name per line": "Elenco nomi di artisti o stili. File '.txt' con un nome per riga",
"Prompt words before artist or style name": "Parole chiave prima del nome dell'artista o dello stile",
"Prompt words after artist or style name": "Parole chiave dopo il nome dell'artista o dello stile",
"Negative Prompt": "Prompt negativo",
"Save": "Salva",
"Send to img2img": "Invia a img2img",
"Send to inpaint": "Invia a Inpaint",
"Send to img2img": "Invia a\nimg2img",
"Send to inpaint": "Invia a\nInpaint",
"Send to extras": "Invia a Extra",
"Make Zip when Save?": "Crea un file ZIP quando si usa 'Salva'",
"Textbox": "Casella di testo",
@ -369,7 +533,7 @@
"Just resize": "Ridimensiona solamente",
"Crop and resize": "Ritaglia e ridimensiona",
"Resize and fill": "Ridimensiona e riempie",
"Advanced loopback": "Advanced loopback",
"Advanced loopback": "Elaborazione ricorsiva avanzata",
"External Image Masking": "Immagine esterna per la mascheratura",
"img2img alternative test": "Test alternativo per img2img",
"img2tiles": "img2tiles",
@ -396,9 +560,9 @@
"Saturation enhancement per image": "Miglioramento della saturazione per ciascuna immagine",
"Use sine denoising strength variation": "Utilizzare la variazione sinusoidale dell'intensità di riduzione del rumore",
"Phase difference": "Differenza di Fase",
"Denoising strength exponentiation": "Esponenziazione dell'intensità di riduzione del rumore",
"Denoising strength exponentiation": "Esponente intensità denoising",
"Use sine zoom variation": "Usa la variazione sinusoidale dello zoom",
"Zoom exponentiation": "Esponeniazione dello Zoom",
"Zoom exponentiation": "Esponente Zoom",
"Use multiple prompts": "Usa prompt multipli",
"Same seed per prompt": "Stesso seme per ogni prompt",
"Same seed for everything": "Stesso seme per tutto",
@ -421,6 +585,7 @@
"Sigma adjustment for finding noise for image": "Regolazione Sigma per trovare il rumore per l'immagine",
"Tile size": "Dimensione piastrella",
"Tile overlap": "Sovrapposizione piastrella",
"New seed for each tile": "Nuovo seme per ogni piastrella",
"alternate img2img imgage": "Immagine alternativa per img2img",
"interpolation values": "Valori di interpolazione",
"Refinement loops": "Cicli di affinamento",
@ -439,7 +604,7 @@
"Reuse seed": "Riusa il seme",
"CFG decay factor": "Fattore di decadimento CFG",
"CFG target": "CFG di destinazione",
"Show/Hide Canvas": "Mostra/Nascondi Tela",
"Show/Hide AlphaCanvas": "Mostra/Nascondi AlphaCanvas",
"Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8": "Impostazioni consigliate: Passi di campionamento: 80-100, Campionatore: Euler a, Intensità denoising: 0.8",
"Pixels to expand": "Pixel da espandere",
"Outpainting direction": "Direzione di Outpainting",
@ -447,13 +612,14 @@
"right": "destra",
"up": "sopra",
"down": "sotto",
"Fall-off exponent (lower=higher detail)": "Esponente di decremento (più basso=maggior dettaglio)",
"Fall-off exponent (lower=higher detail)": "Esponente di decremento (più basso = maggior dettaglio)",
"Color variation": "Variazione di colore",
"Will upscale the image to twice the dimensions; use width and height sliders to set tile size": "Aumenterà l'immagine al doppio delle dimensioni; utilizzare i cursori di larghezza e altezza per impostare la dimensione della piastrella",
"Upscaler": "Ampliamento immagine",
"Lanczos": "Lanczos",
"Nearest": "Nearest",
"LDSR": "LDSR",
"ESRGAN_4x": "ESRGAN_4x",
"BSRGAN": "BSRGAN",
"ScuNET GAN": "ScuNET GAN",
"ScuNET PSNR": "ScuNET PSNR",
"SwinIR 4x": "SwinIR 4x",
@ -509,7 +675,7 @@
"CodeFormer visibility": "Visibilità CodeFormer",
"CodeFormer weight (0 = maximum effect, 1 = minimum effect)": "Peso di CodeFormer (0 = effetto massimo, 1 = effetto minimo)",
"Upscale Before Restoring Faces": "Amplia prima di restaurare i volti",
"Send to txt2img": "Invia a txt2img",
"Send to txt2img": "Invia a\ntxt2img",
"A merger of the two checkpoints will be generated in your": "I due checkpoint verranno fusi nella cartella dei",
"checkpoint": "checkpoint",
"directory.": ".",
@ -535,7 +701,7 @@
"Modules": "Moduli",
"Enter hypernetwork layer structure": "Immettere la struttura del livello della Iperrete",
"Select activation function of hypernetwork": "Selezionare la funzione di attivazione della Iperrete",
"linear": "linear",
"linear": "lineare",
"relu": "relu",
"leakyrelu": "leakyrelu",
"elu": "elu",
@ -563,9 +729,9 @@
"softshrink": "softshrink",
"softsign": "softsign",
"tanhshrink": "tanhshrink",
"threshold": "threshold",
"threshold": "soglia",
"Select Layer weights initialization. relu-like - Kaiming, sigmoid-like - Xavier is recommended": "Seleziona inizializzazione dei pesi dei livelli. relu-like - Kaiming, Si consiglia sigmoid-like - Xavier",
"Normal": "Normal",
"Normal": "Normale",
"KaimingUniform": "KaimingUniform",
"KaimingNormal": "KaimingNormal",
"XavierUniform": "XavierUniform",
@ -609,20 +775,75 @@
"Train Embedding": "Addestra Incorporamento",
"Create an aesthetic embedding out of any number of images": "Crea un'incorporamento estetico da qualsiasi numero di immagini",
"Create images embedding": "Crea incorporamento di immagini",
"Generate Krita Plugin Symlink Command": "Genera comando per creare collegamento simbolico del plugin Krita",
"Launch Krita.": "Avvia Krita.",
"On the menubar, go to": "Sulla barra dei menu, vai a",
"Settings > Manage Resources...": "Settings > Manage Resources...",
"In the window that appears, click": "Nella finestra che appare, fai clic su",
"Open Resource Folder": "Open Resource Folder",
"In the file explorer that appears, look for a folder called": "In Esplora file, una volta aperto, cerca una cartella chiamata",
"pykrita": "pykrita",
"or create it.": "o creala.",
"Enter the": "Entra nella",
"folder and copy the folder location from the address bar.": "cartella e copia il percorso della cartella dalla barra degli indirizzi.",
"Paste the folder location below.": "Incolla il percorso della cartella qui di seguito.",
"Pykrita Folder Location": "Posizione della cartella Pykrita",
"Search for \"Command Prompt\" in the Start Menu, right-click and click \"Run as Administrator...\", paste the follow commands and hit Enter:": "Cerca \"Prompt dei comandi\" nel menu Start, fai clic con il pulsante destro del mouse e fai clic su \"Esegui come amministratore...\", incolla i seguenti comandi e premi Invio:",
"mklink /j \"<path_to_pykrita>\\krita_diff\" \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff\"\nmklink \"<path_to_pykrita>\\krita_diff.desktop\" \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff.desktop\"": "mklink /j \"<path_to_pykrita>\\krita_diff\" \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff\"\nmklink \"<path_to_pykrita>\\krita_diff.desktop\" \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff.desktop\"",
"Linux command:": "comando Linux:",
"ln -s \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff\" \"<path_to_pykrita>/krita_diff\"\nln -s \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff.desktop\" \"<path_to_pykrita>/krita_diff.desktop\"": "ln -s \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff\" \"<path_to_pykrita>/krita_diff\"\nln -s \"C:\\stable-diffusion-webui\\extensions\\auto-sd-paint-ext\\frontends\\krita\\krita_diff.desktop\" \"<path_to_pykrita>/krita_diff.desktop\"",
"NOTE": "NOTE",
": Symlinks will break if you move or rename the repository or any\nof its parent folders or otherwise change the path such that the symlink\nbecomes invalid. In which case, repeat the above steps with the new": ": I collegamenti simbolici si interrompono se si sposta o si rinomina il repository o una qualsiasi\n delle sue cartelle principali o si modifica in altro modo il percorso in modo tale che il collegamento simbolico\ndiventi non valido. In tal caso, ripetere i passaggi precedenti con il nuovo ",
"folder location and (auto-detected) repository location.": "percorso della cartella e percorso del repository (rilevato automaticamente).",
": Ensure": ": Assicurarsi che",
"webui-user.bat": "webui-user.bat",
"/": "/",
"webui-user.sh": "webui-user.sh",
"contains": "contenga",
"--api": "--api",
"in": "in",
"COMMANDLINE_ARGS": "COMMANDLINE_ARGS",
"!": "!",
"Enabling the Krita Plugin": "Abilitazione del plugin Krita",
"Restart Krita.": "Riavvia Krita.",
"Settings > Configure Krita...": "Settings > Configure Krita...",
"On the left sidebar, go to": "Nella barra laterale sinistra, vai a",
"Python Plugin Manager": "Python Plugin Manager",
"Look for": "Cerca",
"Stable Diffusion Plugin": "Stable Diffusion Plugin",
"and tick the checkbox.": "e spunta la casella.",
"Restart Krita again for changes to take effect.": "Riavvia Krita per rendere effettive le modifiche.",
"The": "Il",
"SD Plugin": "SD Plugin",
"docked window should appear on the left of the Krita window. If it does not, look on the menubar under": "dovrebbe apparire come finestra ancorata a sinistra della finestra di Krita. In caso contrario, guarda nella barra dei menu sotto",
"Settings > Dockers": "Settings > Dockers",
"for": "per",
"Next Steps": "Prossimi passi",
"Troubleshooting": "Risoluzione dei problemi",
"Update Guide": "Guida all'aggiornamento",
"Usage Guide": "Guida all'uso",
"TODO: Control/status panel": "TODO: pannello di controllo/stato",
"-1": "-1",
"This extension works well with text captions in comma-separated style (such as the tags generated by DeepBooru interrogator).": "Questa estensione funziona bene con i sottotitoli di testo in stile separato da virgole (come i tag generati dall'interrogatore DeepBooru).",
"Save all changes": "Salva tutte le modifiche",
"Backup original text file (original file will be renamed like filename.000, .001, .002, ...)": "Backup del file di testo originale (il file originale verrà rinominato come nomefile.000, .001, .002, ...)",
"Note:": "Note:",
"New text file will be created if you are using filename as captions.": "Verrà creato un nuovo file di testo se si utilizza il nome del file come didascalia.",
"New text file will be created if you are using filename as captions.": " Verrà creato un nuovo file di testo se si utilizza il nome del file come didascalia.",
"Results": "Risultati",
"Load": "Carica",
"Load from subdirectories": "Carica da sotto cartella",
"Dataset Images": "Immagini del Dataset",
"Displayed Images : 0 / 0 total": "Immagini mostrate : 0 / 0 totali",
"Current Tag Filter :": "Filtro Tag corrente :",
"Current Selection Filter : 0 images": "Filtro di selezione corrente : 0 immagini",
"Selected Image :": "Immagine selezionata :",
"Filter and Edit Tags": "Filtra e modifica i tag",
"Filter by Selection": "Filtra per selezione",
"Edit Caption of Selected Image": "Modifica la didascalia dell'immagine selezionata",
"Search tags / Filter images by tags": "Cerca tag / Filtra le immagini per tag",
"Search Tags": "Cerca tag",
"Clear all filters": "Rimuovi tutti i filtri",
"Clear tag filters": "Cancella i filtri dei tag",
"Clear ALL filters": "Rimuovi tutti i filtri",
"Sort by": "Ordina per",
"Alphabetical Order": "Ordine alfabetico",
"Frequency": "Frequenza",
@ -650,6 +871,13 @@
"ex C.": "esempio C.",
"Original Text = \"A, B, C, D, E\"Selected Tags = \"A, B, D\"Edit Tags = \", X, \"": "Testo originale = \"A, B, C, D, E\"Tag selezionati = \"A, B, D\"Modifica tag = \", X, \"",
"Result = \"X, C, E\"(A->\"\", B->X, D->\"\")": "Risultato = \"X, C, E\"(A->\"\", B->X, D->\"\")",
"Select images from the left gallery.": "Seleziona le immagini dalla galleria di sinistra.",
"Add selection [Enter]": "Aggiungi selezione [Invio]",
"Filter Images": "Filtra le immagini",
"Remove selection [Delete]": "Rimuovi selezione [Canc]",
"Invert selection": "Inverti selezione",
"Clear selection": "Cancella selezione",
"Apply selection filter": "Applica il filtro di selezione",
"Caption of Selected Image": "Didascalia dell'immagine selezionata",
"Copy caption": "Copia didascalia",
"Edit Caption": "Modifica didascalia",
@ -754,7 +982,7 @@
"bilinear": "bilinear",
"nearest": "nearest",
"save_depth_maps": "Salva le mappe di profondità",
"`animation_mode: None` batches on list of *prompts*. (Batch mode disabled atm, only animation_prompts are working)": "`modalità animazione: Nessuno` si inserisce nell'elenco di *prompt*. (Modalità batch disabilitata atm, funzionano solo i prompt di animazione)",
"`animation_mode: None` batches on list of *prompts*. (Batch mode disabled atm, only animation_prompts are working)": "`modalità animazione: Nessuno` si inserisce nell'elenco di *prompt*. (Modalità batch al momento disabilitata, funzionano solo i prompt di animazione)",
"*Important change from vanilla Deforum!*": "*Importante cambiamento rispetto alla versione originale di Deforum!*",
"This script uses the built-in webui weighting settings.": "Questo script utilizza le impostazioni di pesatura webui integrate.",
"So if you want to use math functions as prompt weights,": "Quindi, se vuoi usare le funzioni matematiche come pesi dei prompt,",
@ -764,17 +992,21 @@
"animation_prompts": "Prompt animazione",
"Init settings": "Impostazioni iniziali",
"use_init": "Usa le impostazioni iniziali",
"from_img2img_instead_of_link": "from_img2img_instead_of_link",
"strength_0_no_init": "strength_0_no_init",
"from_img2img_instead_of_link": "da img2img invece che da link",
"strength_0_no_init": "Intensità 0 nessuna inizializzazione",
"strength": "Intensità",
"init_image": "Immagine di inizializzazione",
"use_mask": "Usa maschera",
"use_alpha_as_mask": "Usa alpha come maschera",
"use_alpha_as_mask": "Usa Alpha come maschera",
"invert_mask": "Inverti la maschera",
"overlay_mask": "Sovrapponi la maschera",
"mask_file": "File della maschera",
"mask_contrast_adjust": "regolazione contrasto maschera",
"mask_brightness_adjust": "Regola la luminosità della maschera",
"mask_overlay_blur": "Sfocatura della sovrapposizione della maschera",
"mask_fill": "riempimento maschera",
"full_res_mask": "maschera a piena risoluzione",
"full_res_mask_padding": "riempimento maschera a piena risoluzione",
"Video Input:": "Ingresso video:",
"video_init_path": "Percorso del video di inizializzazione",
"extract_nth_frame": "Estrai ogni ennesimo fotogramma",
@ -791,8 +1023,8 @@
"skip_video_for_run_all": "Salta il video per eseguire tutto",
"fps": "FPS",
"output_format": "Formato di uscita",
"PIL gif": "PIL gif",
"FFMPEG mp4": "FFMPEG mp4",
"PIL gif": "PIL GIF",
"FFMPEG mp4": "FFMPEG MP4",
"ffmpeg_location": "Percorso ffmpeg",
"add_soundtrack": "Aggiungi colonna sonora",
"soundtrack_path": "Percorso colonna sonora",
@ -805,6 +1037,8 @@
"image_path": "Percorso immagine",
"mp4_path": "Percorso MP4",
"Click here after the generation to show the video": "Clicca qui dopo la generazione per mostrare il video",
"Close the video": "Chiudi il video",
"NOTE: If the 'Generate' button doesn't work, go in Settings and click 'Restart Gradio and Refresh...'.": "NOTA: se il pulsante 'Genera' non funziona, vai in Impostazioni e fai clic su 'Riavvia Gradio e Aggiorna...'.",
"Save Settings": "Salva le impostazioni",
"Load Settings": "Carica le impostazioni",
"Path relative to the webui folder." : "Percorso relativo alla cartella webui.",
@ -814,7 +1048,6 @@
"house": "casa",
"portrait": "ritratto",
"spaceship": "nave spaziale",
"anime": "anime",
"cartoon": "cartoon",
"digipa-high-impact": "digipa-high-impact",
"digipa-med-impact": "digipa-med-impact",
@ -916,13 +1149,13 @@
"Generate Info": "Genera Info",
"File Name": "Nome del file",
"Move to favorites": "Aggiungi ai preferiti",
"Renew page": "Aggiorna la pagina",
"Renew Page": "Aggiorna la pagina",
"Number": "Numero",
"set_index": "Imposta indice",
"load_switch": "load_switch",
"turn_page_switch": "turn_page_switch",
"Checkbox": "Casella di controllo",
"Checkbox Group": "Seleziona immagini per",
"load_switch": "Carica",
"turn_page_switch": "Volta pagina",
"Checkbox": "Selezione",
"Checkbox Group": "Seleziona per",
"artists": "Artisti",
"flavors": "Stili",
"mediums": "Tecniche",
@ -932,12 +1165,24 @@
"Abandoned": "Scartati",
"Key word": "Parola chiave",
"Get inspiration": "Ispirami",
"to txt2img": "Invia a txt2img",
"to img2img": "Invia a img2img",
"to txt2img": "Invia a\ntxt2img",
"to img2img": "Invia a\nimg2img",
"Collect": "Salva nei preferiti",
"Don't show again": "Scarta",
"Move out": "Rimuovi",
"set button": "Pulsante imposta",
"set button": "Pulsante imposta",
"Before your text is sent to the neural network, it gets turned into numbers in a process called tokenization. These tokens are how the neural network reads and interprets text. Thanks to our great friends at Shousetsu愛 for inspiration for this feature.": "Prima che il tuo testo venga inviato alla rete neurale, verrà trasformato in numeri in un processo chiamato tokenizzazione. Questi token sono il modo in cui la rete neurale legge e interpreta il testo. Grazie ai nostri grandi amici di Shousetsu愛 per l'ispirazione per questa funzione.",
"Tokenize": "Tokenizzare",
"Text": "Testo",
"Tokens": "Token",
"Video to extract frames from:": "Video da cui estrarre fotogrammi:",
"Only extract keyframes (recommended)": "Estrae solo fotogrammi chiave (consigliato)",
"Extract Frames": "Estrai fotogrammi",
"Extracted Frame Set": "Set di fotogrammi estratti",
"Resize crops to 512x512": "Ridimensiona i ritagli a 512x512",
"Save crops to:": "Salva ritagli in:",
"<": "<",
">": ">",
"Apply settings": "Applica le impostazioni",
"Saving images/grids": "Salva immagini/griglie",
"Always save all generated images": "Salva sempre tutte le immagini generate",
@ -953,6 +1198,8 @@
"Save text information about generation parameters as chunks to png files": "Salva le informazioni di testo dei parametri di generazione come blocchi nel file png",
"Create a text file next to every image with generation parameters.": "Crea un file di testo assieme a ogni immagine con i parametri di generazione.",
"Save a copy of image before doing face restoration.": "Salva una copia dell'immagine prima di eseguire il restauro dei volti.",
"Save a copy of image before applying highres fix.": "Salva una copia dell'immagine prima di applicare la correzione ad alta risoluzione.",
"Save a copy of image before applying color correction to img2img results": "Salva una copia dell'immagine prima di applicare la correzione del colore ai risultati di img2img",
"Quality for saved jpeg images": "Qualità delle immagini salvate in formato JPEG",
"If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG": "Se l'immagine PNG è più grande di 4 MB o qualsiasi dimensione è maggiore di 4000, ridimensiona e salva la copia come JPG",
"Use original name for output filename during batch process in extras tab": "Usa il nome originale per il nome del file di output durante l'elaborazione a lotti nella scheda 'Extra'",
@ -990,16 +1237,19 @@
"Add a second progress bar to the console that shows progress for an entire job.": "Aggiungi una seconda barra di avanzamento alla console che mostra l'avanzamento complessivo del lavoro.",
"Training": "Addestramento",
"Move VAE and CLIP to RAM when training hypernetwork. Saves VRAM.": "Sposta VAE e CLIP nella RAM durante l'addestramento di Iperreti. Risparmia VRAM.",
"Filename word regex": "Filename word regex",
"Filename join string": "Filename join string",
"Saves Optimizer state as separate *.optim file. Training can be resumed with HN itself and matching optim file.": "Salva lo stato dell'ottimizzatore come file *.optim separato. L'addestramento può essere ripreso con lo stesso HN e il file optim corrispondente.",
"Filename word regex": "Espressione regolare per estrarre parole dal nome del file",
"Filename join string": "Stringa per unire le parole estratte dal nome del file",
"Number of repeats for a single input image per epoch; used only for displaying epoch number": "Numero di ripetizioni per una singola immagine di input per epoca; utilizzato solo per visualizzare il numero di epoca",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "Salva un file CSV contenente la perdita nella cartella di registrazione ogni N passaggi, 0 per disabilitare",
"Use cross attention optimizations while training": "Usa le ottimizzazioni di controllo dell'attenzione incrociato durante l'allenamento",
"Stable Diffusion": "Stable Diffusion",
"Checkpoints to cache in RAM": "Checkpoint da memorizzare nella RAM",
"SD VAE": "SD VAE",
"auto": "auto",
"Hypernetwork strength": "Forza della Iperrete",
"Inpainting conditioning mask strength": "Forza della maschera di condizionamento del Inpainting",
"Apply color correction to img2img results to match original colors.": "Applica la correzione del colore ai risultati di img2img in modo che corrispondano ai colori originali.",
"Save a copy of image before applying color correction to img2img results": "Salva una copia dell'immagine prima di applicare la correzione del colore ai risultati di img2img",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "Con img2img, esegue esattamente la quantità di passi specificata dalla barra di scorrimento (normalmente se ne effettuano di meno con meno riduzione del rumore).",
"Enable quantization in K samplers for sharper and cleaner results. This may change existing seeds. Requires restart to apply.": "Abilita la quantizzazione nei campionatori K per risultati più nitidi e puliti. Questo può cambiare i semi esistenti. Richiede il riavvio per applicare la modifica.",
"Emphasis: use (text) to make model pay more attention to text and [text] to make it pay less attention": "Enfasi: utilizzare (testo) per fare in modo che il modello presti maggiore attenzione al testo e [testo] per fargli prestare meno attenzione",
@ -1059,11 +1309,17 @@
"sigma tmin": "sigma tmin",
"sigma noise": "sigma noise",
"Eta noise seed delta": "ETA del delta del seme del rumore",
"Number of columns on image gallery": "Numero di colonne nella galleria di immagini",
"Aesthetic Image Scorer": "Punteggio delle immagini estetiche",
"Save score as EXIF or PNG Info Chunk": "Salva il punteggio come info EXIF o PNG",
"Save score as tag (Windows Only)": "Salva punteggio come etichetta (solo Windows)",
"aesthetic_score": "Punteggio estetico",
"cfg_scale": "Scala CFG",
"sd_model_hash": "Hash del modello SD",
"hash": "Hash",
"Save tags (Windows only)": "Salva etichette (solo Windows)",
"Save category (Windows only)": "Salva categoria (solo Windows)",
"Save generation params text": "Salva testo parametri di generazione",
"Force CPU (Requires Custom Script Reload)": "Forza CPU (richiede il ricaricamento dello script personalizzato)",
"Number of columns on image gallery": "Numero di colonne nella galleria di immagini",
"Images Browser": "Galleria immagini",
"Preload images at startup": "Precarica le immagini all'avvio",
"Number of columns on the page": "Numero di colonne nella pagina",
@ -1076,6 +1332,7 @@
"Reload custom script bodies (No ui updates, No restart)": "Ricarica gli script personalizzati (nessun aggiornamento dell'interfaccia utente, nessun riavvio)",
"Restart Gradio and Refresh components (Custom Scripts, ui.py, js and css only)": "Riavvia Gradio e aggiorna i componenti (solo script personalizzati, ui.py, js e css)",
"Installed": "Installato",
"Available": "Disponibile",
"Install from URL": "Installa da URL",
"Apply and restart UI": "Applica e riavvia l'interfaccia utente",
"Check for updates": "Controlla aggiornamenti",
@ -1085,22 +1342,36 @@
"aesthetic-gradients": "Gradienti Estetici (CLIP)",
"https://github.com/AUTOMATIC1111/stable-diffusion-webui-aesthetic-gradients": "https://github.com/AUTOMATIC1111/stable-diffusion-webui-aesthetic-gradients",
"unknown": "sconosciuto",
"aesthetic-image-scorer": "Punteggio delle immagini estetiche",
"https://github.com/tsngo/stable-diffusion-webui-aesthetic-image-scorer": "https://github.com/tsngo/stable-diffusion-webui-aesthetic-image-scorer",
"auto-sd-paint-ext": "KRITA Plugin",
"https://github.com/Interpause/auto-sd-paint-ext": "https://github.com/Interpause/auto-sd-paint-ext",
"dataset-tag-editor": "Dataset Tag Editor",
"https://github.com/toshiaki1729/stable-diffusion-webui-dataset-tag-editor.git": "https://github.com/toshiaki1729/stable-diffusion-webui-dataset-tag-editor.git",
"deforum": "Deforum",
"https://github.com/deforum-art/deforum-for-automatic1111-webui/": "https://github.com/deforum-art/deforum-for-automatic1111-webui/",
"deforum-for-automatic1111-webui": "Deforum",
"https://github.com/deforum-art/deforum-for-automatic1111-webui": "https://github.com/deforum-art/deforum-for-automatic1111-webui",
"novelai-2-local-prompt": "Converti NovelAI",
"https://github.com/animerl/novelai-2-local-prompt": "https://github.com/animerl/novelai-2-local-prompt",
"sd-dynamic-prompts": "Prompt dinamici",
"https://github.com/adieyal/sd-dynamic-prompts": "https://github.com/adieyal/sd-dynamic-prompts",
"stable-diffusion-webui-aesthetic-image-scorer": "Punteggio immagini estetiche",
"https://github.com/tsngo/stable-diffusion-webui-aesthetic-image-scorer": "https://github.com/tsngo/stable-diffusion-webui-aesthetic-image-scorer",
"stable-diffusion-webui-artists-to-study": "Artisti per studiare",
"https://github.com/camenduru/stable-diffusion-webui-artists-to-study": "https://github.com/camenduru/stable-diffusion-webui-artists-to-study",
"stable-diffusion-webui-images-browser": "Galleria immagini",
"https://github.com/yfszzx/stable-diffusion-webui-images-browser": "https://github.com/yfszzx/stable-diffusion-webui-images-browser",
"stable-diffusion-webui-inspiration": "Ispirazione",
"https://github.com/yfszzx/stable-diffusion-webui-inspiration": "https://github.com/yfszzx/stable-diffusion-webui-inspiration",
"stable-diffusion-webui-tokenizer": "Tokenizzatore",
"https://github.com/AUTOMATIC1111/stable-diffusion-webui-tokenizer.git": "https://github.com/AUTOMATIC1111/stable-diffusion-webui-tokenizer.git",
"tag-autocomplete": "Autocompletamento etichette",
"https://github.com/DominikDoom/a1111-sd-webui-tagcomplete.git": "https://github.com/DominikDoom/a1111-sd-webui-tagcomplete.git",
"training-picker": "Selezionatore per l'addestramento",
"https://github.com/Maurdekye/training-picker.git": "https://github.com/Maurdekye/training-picker.git",
"unprompted": "Unprompted",
"https://github.com/ThereforeGames/unprompted": "https://github.com/ThereforeGames/unprompted",
"wildcards": "Termini Jolly",
"https://github.com/AUTOMATIC1111/stable-diffusion-webui-wildcards.git": "https://github.com/AUTOMATIC1111/stable-diffusion-webui-wildcards.git",
"Load from:": "Carica da:",
"Extension index URL": "URL dell'indice delle Estensioni",
"URL for extension's git repository": "URL del repository GIT dell'estensione",
"Local directory name": "Nome cartella locale",
"Install": "Installa",
@ -1131,7 +1402,8 @@
"Make an attempt to produce a picture similar to what would have been produced with same seed at specified resolution": "Prova a produrre un'immagine simile a quella che sarebbe stata prodotta con lo stesso seme alla risoluzione specificata",
"This text is used to rotate the feature space of the imgs embs": "Questo testo viene utilizzato per ruotare lo spazio delle funzioni delle immagini incorporate",
"How many times to repeat processing an image and using it as input for the next iteration": "Quante volte ripetere l'elaborazione di un'immagine e utilizzarla come input per l'iterazione successiva",
"Hello, StylePile here.\nUntil some weird bug gets fixed you will see this even if the script itself is not active. Meanwhile, some hints to take your artwork to new heights:\nUse the 'Focus on' dropdown to select complex presets. Toggle selections below (with or without Focus) to affect your results. Mix and match to get some interesting results. \nAnd some general Stable Diffusion tips that will take your designs to next level:\nYou can add parenthesis to make parts of the prompt stronger. So (((cute))) kitten will make it extra cute (try it out). This is alsow important if a style is affecting your original prompt too much. Make that prompt stronger by adding parenthesis around it, like this: ((promt)).\nYou can type promts like [A|B] to sequentially use terms one after another on each step. So, like [cat|dog] will produce a hybrid catdog. And [A:B:0.4] to switch to other terms after the first one has been active for a certain percentage of steps. So [cat:dog:0.4] will build a cat 40% of the time and then start turning it into a dog. This needs more steps to work properly.": "Salve, qui è StylePile.\nFinché qualche strano bug non verrà risolto, vedrai questo testo anche se lo script non è attivo. Nel frattempo, alcuni suggerimenti per portare la tua grafica a nuovi livelli:\nUtilizza il menu a discesa 'Focus on' per selezionare valori predefiniti complessi. Attiva o disattiva le selezioni seguenti (con o senza Focus) per influire sui risultati. Mescola e abbina per ottenere risultati interessanti. \nE alcuni suggerimenti generali su Stable Diffusion che porteranno i tuoi risultati a un livello superiore:\nPuoi aggiungere parentesi per aumentare l'influenza di certe parti del prompt. Quindi '(((cute))) kitten' lo renderà molto carino (fai delle prove). Questo è importante quando uno stile influisce troppo sul prompt originale. Rendi più forte quel prompt aggiungendo delle parentesi intorno ad esso, così: ((promt)).\nPuoi digitare prompt nel formato [A|B] per usare in sequenza i termini uno dopo l'altro in ogni passaggio. Quindi, come [cat|dog] produrrà un 'canegatto' ibrido. E [A:B:0.4] per passare ad altri termini dopo che il primo è stato attivo per una certa percentuale di passaggi. Quindi [cat:dog:0.4] genererà un gatto il 40% dei passaggi e poi inizierà a trasformarlo in un cane. Sono richiesti più passaggi perchè funzioni correttamente.",
"Insert [X] anywhere in main prompt to sequentially insert values from here.": "Inserisci [X] ovunque nel prompt principale per inserire in sequenza i valori da qui.",
"Insert [R] anywhere in main prompt to randomly insert values from here.": "Inserisci [R] ovunque nel prompt principale per inserire i valori in modo casuale da qui.",
"Enter one prompt per line. Blank lines will be ignored.": "Immettere un prompt per riga. Le righe vuote verranno ignorate.",
"Separate values for X axis using commas.": "Separare i valori per l'asse X usando le virgole.",
"Separate values for Y axis using commas.": "Separare i valori per l'asse Y usando le virgole.",
@ -1158,8 +1430,10 @@
"1st and last digit must be 1. ex:'1, 2, 1'": "La prima e l'ultima cifra devono essere 1. Es.:'1, 2, 1'",
"Path to directory with input images": "Percorso della cartella con immagini di input",
"Path to directory where to write outputs": "Percorso della cartella in cui scrivere i risultati",
"C:\\\\...\\pykrita": "C:\\\\...\\pykrita",
"C:\\directory\\of\\datasets": "C:\\cartella\\del\\dataset",
"Input images directory": "Cartella di input delle immagini",
"Prompt for tokenization": "Prompt for tokenization",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; leave empty for default.": "Usa i seguenti tag per definire come vengono scelti i nomi dei file per le immagini: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed ], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; lasciare vuoto per usare l'impostazione predefinita.",
"If this option is enabled, watermark will not be added to created images. Warning: if you do not add watermark, you may be behaving in an unethical manner.": "Se questa opzione è abilitata, la filigrana non verrà aggiunta alle immagini create. Attenzione: se non aggiungi la filigrana, potresti comportarti in modo non etico.",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; leave empty for default.": "Utilizzare i seguenti tag per definire come vengono scelte le sottodirectory per le immagini e le griglie: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; lasciare vuoto per usare l'impostazione predefinita.",
@ -1169,5 +1443,123 @@
"Only applies to inpainting models. Determines how strongly to mask off the original image for inpainting and img2img. 1.0 means fully masked, which is the default behaviour. 0.0 means a fully unmasked conditioning. Lower values will help preserve the overall composition of the image, but will struggle with large changes.": "Si applica solo ai modelli di pittura. Determina con quale forza mascherare l'immagine originale per inpainting e img2img. 1.0 significa completamente mascherato, che è il comportamento predefinito. 0.0 significa un condizionamento completamente non mascherato. Valori più bassi aiuteranno a preservare la composizione generale dell'immagine, ma avranno difficoltà con grandi cambiamenti.",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "Elenco dei nomi delle impostazioni, separati da virgole, per le impostazioni che dovrebbero essere visualizzate nella barra di accesso rapido in alto, anziché nella normale scheda delle impostazioni. Vedi modules/shared.py per impostare i nomi. Richiede il riavvio per applicare.",
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "Se questo valore è diverso da zero, verrà aggiunto al seed e utilizzato per inizializzare il generatore di numeri casuali per il rumore quando si utilizzano campionatori con ETA. Puoi usarlo per produrre ancora più variazioni di immagini, oppure puoi usarlo per abbinare le immagini di altri software se sai cosa stai facendo.",
"Leave empty for auto": "Lasciare vuoto per automatico"
}
"Leave empty for auto": "Lasciare vuoto per automatico",
"NAIConvert": "Converti\nNovelAI",
"History": "Crononolgia",
"specify a custom settings file and ignore settings displayed in the interface": "specificare un file di impostazioni personalizzate e ignorare le impostazioni visualizzate nell'interfaccia",
"the path to a custom settings file": "il percorso di un file delle impostazioni personalizzate",
"The width of the output images, in pixels (must be a multiple of 64)": "La larghezza delle immagini di output, in pixel (deve essere un multiplo di 64)",
"The height of the output images, in pixels (must be a multiple of 64)": "L'altezza delle immagini di output, in pixel (deve essere un multiplo di 64)",
"enable additional seed settings": "abilitare impostazioni seme aggiuntive",
"How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results": "Quante volte migliorare l'immagine generata in modo iterativo; valori più alti richiedono più tempo; valori molto bassi possono produrre risultati negativi",
"output images will be placed in a folder with this name, inside of the img2img output folder": "le immagini di output verranno collocate in una cartella con questo nome, all'interno della cartella di output img2img",
"specify the format of the filename for output images": "specificare il formato del nome file per le immagini di output",
"defines the seed behavior that is used for animations": "definisce il comportamento del seme utilizzato per le animazioni",
"the seed value will increment by 1 for each subsequent frame of the animation": "il valore del seme aumenterà di 1 per ogni fotogramma successivo dell'animazione",
"selects the type of animation": "seleziona il tipo di animazione",
"only 2D motion parameters will be used, but this mode uses the least amount of VRAM. You can optionally enable flip_2d_perspective to enable some psuedo-3d animation parameters while in 2D mode.": "verranno utilizzati solo i parametri di movimento 2D, ma questa modalità utilizza la quantità minima di VRAM. Puoi facoltativamente abilitare flip_2d_perspective per abilitare alcuni parametri di animazione psuedo-3d mentre sei in modalità 2D.",
"the maximum number of output images to be created": "il numero massimo di immagini di output da creare",
"controls handling method of pixels to be generated when the image is smaller than the frame.": "controlla il metodo di gestione dei pixel da generare quando l'immagine è più piccola del fotogramma.",
"repeats the edge of the pixels, and extends them. Animations with quick motion may yield lines where this border function was attempting to populate pixels into the empty space created.": "ripete il bordo dei pixel e li estende. Le animazioni con movimento rapido possono produrre linee in cui questa funzione di bordo stava tentando di popolare i pixel nello spazio vuoto creato.",
"2D operator to rotate canvas clockwise/anticlockwise in degrees per frame": "Operatore 2D per ruotare la tela in senso orario/antiorario in gradi per fotogramma",
"2D operator that scales the canvas size, multiplicatively. [static = 1.0]": "Operatore 2D che ridimensiona le dimensioni della tela, in modo moltiplicativo. [statico = 1.0]",
"2D & 3D operator to move canvas left/right in pixels per frame": "Operatore 2D e 3D per spostare la tela a sinistra/destra in pixel per fotogramma",
"2D & 3D operator to move canvas up/down in pixels per frame": "Operatore 2D e 3D per spostare la tela su/giù in pixel per fotogramma",
"3D operator to move canvas towards/away from view [speed set by FOV]": "Operatore 3D per spostare la tela da/verso la visuale [velocità impostata da FOV]",
"3D operator to tilt canvas up/down in degrees per frame": "Operatore 3D per inclinare la tela su/giù in gradi per fotogramma",
"3D operator to pan canvas left/right in degrees per frame": "Operatore 3D per eseguire una panoramica della tela a sinistra/destra in gradi per fotogramma",
"3D operator to roll canvas clockwise/anticlockwise": "Operatore 3D per ruotare la tela in senso orario/antiorario",
"enables 2D mode functions to simulate faux 3D movement": "abilita le funzioni della modalità 2D per simulare falsi movimenti 3D",
"the roll effect angle": "l'angolo dell'effetto di rollio",
"the tilt effect angle": "l'angolo dell'effetto di inclinazione",
"the pan effect angle": "l'angolo dell'effetto panoramica",
"the 2D vanishing point of perspective (recommended range 30-160)": "il punto di fuga prospettico 2D (intervallo consigliato 30-160)",
"amount of graininess to add per frame for diffusion diversity": "quantità di granulosità da aggiungere per fotogramma per la diversità di diffusione",
"amount of presence of previous frame to influence next frame, also controls steps in the following formula [steps - (strength_schedule * steps)]": "quantità di presenza del fotogramma precedente per influenzare il fotogramma successivo, controlla anche i passaggi nella seguente formula [passi - (forza pianificazione * steps)]",
"adjusts the overall contrast per frame [default neutral at 1.0]": "regola il contrasto generale per fotogramma [neutro predefinito a 1.0]",
"how closely the image should conform to the prompt. Lower values produce more creative results. (recommended range 5-15)": "quanto l'immagine deve essere conforme al prompt. Valori più bassi producono risultati più creativi. (intervallo consigliato 5-15)",
"adjusts the scale at which the canvas is moved in 3D by the translation_z value. [maximum range -180 to +180, with 0 being undefined. Values closer to 180 will make the image have less depth, while values closer to 0 will allow more depth]": "regola la scala alla quale la tela viene spostata in 3D in base al valore traslazione_z. [intervallo massimo da -180 a +180, con 0 non definito. Valori più vicini a 180 renderanno l'immagine meno profonda, mentre valori più vicini a 0 consentiranno una maggiore profondità]",
"allows you to specify seeds at a specific schedule, if seed_behavior is set to schedule.": "ti consente di specificare i semi in base a una pianificazione specifica, se seed_behavior è impostato su pianificazione.",
"The color coherence will attempt to sample the overall pixel color information, and trend those values analyzed in the first frame to be applied to future frames.": "La coerenza del colore tenterà di campionare le informazioni complessive sul colore dei pixel e tendere quei valori analizzati nel primo fotogramma da applicare ai fotogrammi futuri.",
"LAB is a more linear approach to mimic human perception of color space - a good default setting for most users.": "LAB è un approccio più lineare per imitare la percezione umana dello spazio colore, una buona impostazione predefinita per la maggior parte degli utenti.",
"The default setting of 1 will cause every frame to receive diffusion in the sequence of image outputs. A setting of 2 will only diffuse on every other frame, yet motion will still be in effect. The output of images during the cadence sequence will be automatically blended, additively and saved to the specified drive. This may improve the illusion of coherence in some workflows as the content and context of an image will not change or diffuse during frames that were skipped. Higher values of 4-8 cadence will skip over a larger amount of frames and only diffuse the “Nth” frame as set by the diffusion_cadence value. This may produce more continuity in an animation, at the cost of little opportunity to add more diffused content. In extreme examples, motion within a frame will fail to produce diverse prompt context, and the space will be filled with lines or approximations of content - resulting in unexpected animation patterns and artifacts. Video Input & Interpolation modes are not affected by diffusion_cadence.": "L'impostazione predefinita di 1 farà sì che ogni fotogramma riceva la diffusione nella sequenza di output delle immagini. Un'impostazione di 2 si diffonderà solo su ogni altro fotogramma, ma il movimento sarà comunque attivo. L'output delle immagini durante la sequenza di cadenza verrà automaticamente miscelato, additivato e salvato nell'unità specificata. Ciò può migliorare l'illusione di coerenza in alcuni flussi di lavoro poiché il contenuto e il contesto di un'immagine non cambiano o si diffondono durante i fotogrammi che sono stati ignorati. Valori più alti di 4-8 cadenza salteranno una quantità maggiore di fotogrammi e diffonderanno solo il fotogramma 'Nth' come impostato dal valore di diffusione_cadenza. Ciò può produrre più continuità in un'animazione, a costo di poche opportunità di aggiungere contenuti più diffusi. In esempi estremi, il movimento all'interno di un fotogramma non riuscirà a produrre un contesto di richiesta diverso e lo spazio sarà riempito con linee o approssimazioni di contenuto, risultando in modelli di animazione e artefatti inaspettati. Le modalità di ingresso video e interpolazione non sono influenzate da diffusion_cadence.",
"enables instructions to warp an image dynamically in 3D mode only.": "abilita le istruzioni per deformare un'immagine dinamicamente solo in modalità 3D.",
"sets a midpoint at which a depthmap is to be drawn: range [-1 to +1]": "imposta un punto medio in cui deve essere disegnata una mappa di profondità: range [da -1 a +1]",
"instructs the handling of pixels outside the field of view as they come into the scene.": "indica la gestione dei pixel al di fuori del campo visivo quando entrano nella scena.",
"choose from Bicubic, Bilinear or Nearest modes. (Recommended: Bicubic)": "scegliere tra le modalità Bicubico, Bilineare o Più vicino. (Consigliato: Bicubico)",
"will output a greyscale depth map image alongside the output images.": "produrrà un'immagine della mappa di profondità in scala di grigi insieme alle immagini di output.",
"Diffuse the first frame based on an image, similar to img2img.": "Diffondi il primo fotogramma in base a un'immagine, simile a img2img.",
"Set the strength to 0 automatically when no init image is used": "Imposta automaticamente l'intensità su 0 quando non viene utilizzata alcuna immagine iniziale",
"Controls the strength of the diffusion on the init image. 0 = disabled": "Controlla l'intensità della diffusione sull'immagine iniziale 0 = disabilitato",
"the path to your init image": "il percorso alla tua immagine iniziale",
"Use a grayscale image as a mask on your init image. Whiter areas of the mask are areas that change more.": "Usa un'immagine in scala di grigi come maschera sull'immagine iniziale. Le aree più bianche della maschera sono aree che cambiano di più.",
"use the alpha channel of the init image as the mask": "usa il canale alfa dell'immagine iniziale come maschera",
"Inverts the colors of the mask": "Inverte i colori della maschera",
"Overlay the masked image at the end of the generation so it does not get degraded by encoding and decoding": "Sovrapponi l'immagine mascherata alla fine della generazione in modo che non venga degradata dalla codifica e dalla decodifica",
"the path to your mask image": "il percorso per l'immagine della maschera",
"adjust the brightness of the mask. Should be a positive number, with 1.0 meaning no adjustment.": "regola la luminosità della maschera. Dovrebbe essere un numero positivo, con 1.0 che significa nessuna correzione.",
"Blur edges of final overlay mask, if used. Minimum = 0 (no blur)": "Sfoca i bordi della maschera di sovrapposizione finale, se utilizzata. Minimo = 0 (nessuna sfocatura)",
"the directory at which your video file is located for Video Input mode only.": "la cartella in cui si trova il file video per la modalità solo Ingresso video.",
"during the run sequence, only frames specified by this value will be extracted, saved, and diffused upon. A value of 1 indicates that every frame is to be accounted for. Values of 2 will use every other frame for the sequence. Higher values will skip that number of frames respectively.": "durante la sequenza di esecuzione, verranno estratti, salvati e diffusi solo i frame specificati da questo valore. Un valore di 1 indica che deve essere contabilizzato ogni frame. I valori di 2 utilizzeranno ogni altro fotogramma per la sequenza. Valori più alti salteranno rispettivamente quel numero di fotogrammi.",
"when enabled, will re-extract video frames each run. When using video_input mode, the run will be instructed to write video frames to the drive. If youve already populated the frames needed, uncheck this box to skip past redundant extraction, and immediately start the render. If you have not extracted frames, you must run at least once with this box checked to write the necessary frames.": "quando abilitato, estrarrà nuovamente i fotogrammi video ad ogni esecuzione. Quando si utilizza la modalità Ingresso video, all'esecuzione verrà richiesto di scrivere fotogrammi video sull'unità. Se avete già popolato i fotogrammi necessari, deselezionate questa casella per saltare l'estrazione ridondante e avviare immediatamente il rendering. Se non sono stati estratti i frame, è necessario eseguire almeno una volta l'operazione con questa casella selezionata per scrivere i fotogrammi necessari.",
"video_input mode only, enables the extraction and use of a separate video file intended for use as a mask. White areas of the extracted video frames will not be affected by diffusion, while black areas will be fully effected. Lighter/darker areas are affected dynamically.": "in modalità solo video Input, consente l'estrazione e l'utilizzo di un file video separato destinato all'uso come maschera. Le aree bianche dei fotogrammi video estratti non saranno interessate dalla diffusione, mentre le aree nere saranno interamente interessate. Le aree più chiare/scure sono interessate in modo dinamico.",
"the directory in which your mask video is located.": "la cartella in cui si trova il video della maschera.",
"selects whether to ignore prompt schedule or _x_frames.": "seleziona se ignorare la pianificazione dei prompt o _x_frames.",
"the number of frames to transition thru between prompts (when interpolate_key_frames = true, then the numbers in front of the animation prompts will dynamically guide the images based on their value. If set to false, will ignore the prompt numbers and force interpole_x_frames value regardless of prompt number)": "il numero di fotogrammi per la transizione tra i prompt (quando interpola fotogrammi chiave = true, i numeri davanti ai prompt dell'animazione guideranno dinamicamente le immagini in base al loro valore. Se impostato su false, ignorerà i numeri dei prompt e forzerà il valore di interpola x fotogrammi indipendentemente da numero di richiesta)",
"instructs the run to start from a specified point": "avvia l'esecuzione da un punto specificato",
"the required timestamp to reference when resuming. Currently only available in 2D & 3D mode, the timestamp is saved as the settings .txt file name as well as images produced during your previous run. The format follows: yyyymmddhhmmss - a timestamp of when the run was started to diffuse.": "l'orario a cui fare riferimento quando si riprende. Attualmente disponibile solo in modalità 2D e 3D, l'orario viene salvato come nome del file .txt delle impostazioni e come immagini prodotte durante la corsa precedente. Il formato segue: yyyymmddhhmmss - l'orario di quando è stata avviata la diffusione dell'esecuzione.",
"when checked, do not output a video": "se selezionato, non produce un video",
"The frames per second that the video will run at": "I fotogrammi al secondo a cui verrà eseguito il video",
"select the type of video file to output": "selezionare il tipo di file video da produrre",
"create an animated GIF": "crea una GIF animata",
"the path to where ffmpeg is located": "il percorso in cui si trova ffmpeg",
"when this box is checked, and FFMPEG mp4 is selected as the output format, an audio file will be multiplexed with the video.": "quando questa casella è selezionata e FFMPEG mp4 è selezionato come formato di output, un file audio verrà multiplexato con il video.",
"the path to an audio file to accompany the video": "tIl percorso di un file audio per accompagnare il video",
"when this is unchecked, the video will automatically be created in the same output folder as the images. Check this box to specify different settings for the creation of the video, specified by the following options": "quando questo è deselezionato, il video verrà creato automaticamente nella stessa cartella di output delle immagini. Selezionare questa casella per specificare impostazioni diverse per la creazione del video, specificate dalle seguenti opzioni",
"render each step of diffusion as a separate frame": "renderizzare ogni fase di diffusione come fotogramma separato",
"the maximum number of frames to include in the video, when use_manual_settings is checked": "il numero massimo di fotogrammi da includere nel video, quando è selezionato 'Usa impostazioni manuali'",
"the location of images to create the video from, when use_manual_settings is checked": "la posizione delle immagini da cui creare il video, quando è selezionato 'Usa impostazioni manuali'",
"the output location of the mp4 file, when use_manual_settings is checked": "il percorso di output del file mp4, quando è selezionato 'Usa impostazioni manuali'",
"Autocomplete options": "Opzioni di autocompletamento",
"Enable Autocomplete": "Abilita autocompletamento",
"Append commas": "Aggiungi virgole",
"Error": "Errore",
"F": "F",
"Time taken:": "Tempo impiegato:",
"latest": "aggiornato",
"behind": "da aggiornare",
"AlphaCanvas": "AlphaCanvas",
"Hue:0": "Hue:0",
"S:0": "S:0",
"L:0": "L:0",
"Load Canvas": "Carica Canvas",
"Save Canvas": "Salva Canvas",
"Close": "Chiudi",
"Grab Results": "Ottieni risultati",
"Apply Patch": "Applica Patch",
"Description": "Descrizione",
"Action": "Azione",
"Aesthetic Gradients": "Gradienti estetici",
"Create an embedding from one or few pictures and use it to apply their style to generated images.": "Crea un incorporamento da una o poche immagini e lo usa per applicare il loro stile alle immagini generate.",
"Sample extension. Allows you to use __name__ syntax in your prompt to get a random line from a file named name.txt in the wildcards directory. Also see Dynamic Prompts for similar functionality.": "Estensione del campione. Consente di utilizzare la sintassi __name__ nel prompt per ottenere una riga casuale da un file denominato name.txt nella cartella dei termini jolly. Vedi anche 'Prompt dinamici' per funzionalità simili.",
"Dynamic Prompts": "Prompt dinamici",
"Implements an expressive template language for random or combinatorial prompt generation along with features to support deep wildcard directory structures.": "Implementa un modello di linguaggio espressivo per la generazione di prompt casuale o combinatoria insieme a funzionalità per supportare cartelle strutturate contenenti termini jolly.",
"Image browser": "Galleria immagini",
"Provides an interface to browse created images in the web browser.": "Fornisce un'interfaccia nel browser web per sfogliare le immagini create.",
"Randomly display the pictures of the artist's or artistic genres typical style, more pictures of this artist or genre is displayed after selecting. So you don't have to worry about how hard it is to choose the right style of art when you create.": "Visualizza in modo casuale le immagini dello stile tipico dell'artista o dei generi artistici, dopo la selezione vengono visualizzate più immagini di questo artista o genere. Così non dovete preoccuparvi della difficoltà di scegliere lo stile artistico giusto quando create.",
"The official port of Deforum, an extensive script for 2D and 3D animations, supporting keyframable sequences, dynamic math parameters (even inside the prompts), dynamic masking, depth estimation and warping.": "Il porting ufficiale di Deforum, uno script completo per animazioni 2D e 3D, che supporta sequenze di fotogrammi chiave, parametri matematici dinamici (anche all'interno dei prompt), mascheramento dinamico, stima della profondità e warping.",
"Artists to study": "Artisti per studiare",
"Shows a gallery of generated pictures by artists separated into categories.": "Mostra una galleria di immagini generate in base agli artisti e suddivise in categorie.",
"Calculates aesthetic score for generated images using CLIP+MLP Aesthetic Score Predictor based on Chad Scorer": "Calcola il punteggio estetico per le immagini generate utilizzando il predittore del punteggio estetico CLIP+MLP basato su Chad Scorer",
"Lets you edit captions in training datasets.": "Consente di modificare le didascalie nei set di dati di addestramento.",
"Krita Plugin.": "Krita Plugin.",
"Adds a tab to the webui that allows the user to automatically extract keyframes from video, and manually extract 512x512 crops of those frames for use in model training.": "Aggiunge una scheda all'interfaccia Web che consente all'utente di estrarre automaticamente i fotogrammi chiave dal video ed estrarre manualmente ritagli 512x512 di quei fotogrammi da utilizzare nell'addestramento del modello.",
"Allows you to include various shortcodes in your prompts. You can pull text from files, set up your own variables, process text through conditional functions, and so much more - it's like wildcards on steroids..": "Consente di includere vari codici brevi nei prompt. Puoi estrarre il testo dai file, impostare le tue variabili, elaborare il testo tramite funzioni condizionali e molto altro ancora: è come i termini jolly ma potenziato.",
"Booru tag autocompletion": "Completamento automatico del tag Booru",
"Displays autocompletion hints for tags from image booru boards such as Danbooru. Uses local tag CSV files and includes a config for customization.": "Visualizza suggerimenti per il completamento automatico dei tag dalle schede di immagini Booru, come Danbooru. Utilizza file CSV di tag locali e include una configurazione per la personalizzazione.",
"Add a button to convert the prompts used in NovelAI for use in the WebUI. In addition, add a button that allows you to recall a previously used prompt.": "Aggiunge un pulsante per convertire i prompt utilizzati in NovelAI per l'uso in Stable Diffusion Web UI. Inoltre, aggiunge un pulsante che permette di richiamare un prompt utilizzato in precedenza.",
"tokenizer": "Tokenizzatore",
"Adds a tab that lets you preview how CLIP model would tokenize your text.": "Aggiunge una scheda che consente di visualizzare in anteprima il modo in cui il modello CLIP tokenizzerebbe il vostro testo.",
"parameters": "Parametri",
"The directory does not exist": "La cartella non esiste"
}

View file

@ -16,6 +16,7 @@
"A merger of the two checkpoints will be generated in your": "체크포인트들이 병합된 결과물이 당신의",
"A value that determines the output of random number generator - if you create an image with same parameters and seed as another image, you'll get the same result": "난수 생성기의 결과물을 지정하는 값 - 동일한 설정값과 동일한 시드를 적용 시, 완전히 똑같은 결과물을 얻게 됩니다.",
"Action": "작업",
"Add a button to convert the prompts used in NovelAI for use in the WebUI. In addition, add a button that allows you to recall a previously used prompt.": "NovelAI에서 사용되는 프롬프트를 WebUI에서 사용할 수 있게 변환하는 버튼을 추가합니다. 덤으로 이전에 사용한 프롬프트를 불러오는 버튼도 추가됩니다.",
"Add a random artist to the prompt.": "프롬프트에 랜덤한 작가 추가",
"Add a second progress bar to the console that shows progress for an entire job.": "콘솔에 전체 작업의 진행도를 보여주는 2번째 프로그레스 바 추가하기",
"Add difference": "차이점 추가",
@ -24,6 +25,8 @@
"Add model hash to generation information": "생성 정보에 모델 해시 추가",
"Add model name to generation information": "생성 정보에 모델 이름 추가",
"Add number to filename when saving": "이미지를 저장할 때 파일명에 숫자 추가하기",
"Adds a tab that lets you preview how CLIP model would tokenize your text.": "CLIP 모델이 텍스트를 어떻게 토큰화할지 미리 보여주는 탭을 추가합니다.",
"Adds a tab to the webui that allows the user to automatically extract keyframes from video, and manually extract 512x512 crops of those frames for use in model training.": "WebUI에 비디오로부터 자동으로 키프레임을 추출하고, 그 키프레임으로부터 모델 훈련에 사용될 512x512 이미지를 잘라낼 수 있는 탭을 추가합니다.",
"Aesthetic Gradients": "스타일 그라디언트",
"Aesthetic Image Scorer": "스타일 이미지 스코어러",
"Aesthetic imgs embedding": "스타일 이미지 임베딩",
@ -32,6 +35,7 @@
"Aesthetic text for imgs": "스타일 텍스트",
"Aesthetic weight": "스타일 가중치",
"Allowed categories for random artists selection when using the Roll button": "랜덤 버튼을 눌러 무작위 작가를 선택할 때 허용된 카테고리",
"Allows you to include various shortcodes in your prompts. You can pull text from files, set up your own variables, process text through conditional functions, and so much more - it's like wildcards on steroids.": "프롬프트에 다양한 숏코드를 추가할 수 있게 해줍니다. 파일로부터 텍스트 추출, 변수 설정, 조건 함수로 텍스트 처리 등등 - 스테로이드를 맞은 와일드카드라 할 수 있죠.",
"Always print all generation info to standard output": "기본 아웃풋에 모든 생성 정보 항상 출력하기",
"Always save all generated image grids": "생성된 이미지 그리드 항상 저장하기",
"Always save all generated images": "생성된 이미지 항상 저장하기",
@ -43,6 +47,7 @@
"Apply selected styles to current prompt": "현재 프롬프트에 선택된 스타일 적용",
"Apply settings": "설정 적용하기",
"Artists to study": "연구할만한 작가들",
"auto": "자동",
"Auto focal point crop": "초점 기준 크롭(자동 감지)",
"Autocomplete options": "자동완성 설정",
"Available": "지원되는 확장기능 목록",
@ -52,6 +57,7 @@
"Batch Process": "이미지 여러장 처리",
"Batch size": "배치 크기",
"behind": "최신 아님",
"Booru tag autocompletion": "Booru 태그 자동완성",
"BSRGAN 4x": "BSRGAN 4x",
"built with gradio": "gradio로 제작되었습니다",
"Calculates aesthetic score for generated images using CLIP+MLP Aesthetic Score Predictor based on Chad Scorer": "Chad 스코어러를 기반으로 한 CLIP+MLP 스타일 점수 예측기를 이용해 생성된 이미지의 스타일 점수를 계산합니다.",
@ -112,6 +118,7 @@
"Directory for saving images using the Save button": "저장 버튼을 이용해 저장하는 이미지들의 저장 경로",
"Directory name pattern": "디렉토리명 패턴",
"directory.": "저장 경로에 저장됩니다.",
"Displays autocompletion hints for tags from image booru boards such as Danbooru. Uses local tag CSV files and includes a config for customization.": "Danbooru 같은 이미지 booru 보드의 태그에 대한 자동완성 힌트를 보여줍니다. 로컬 환경에 저장된 CSV 파일을 사용하고 조정 가능한 설정 파일이 포함되어 있습니다.",
"Do not add watermark to images": "이미지에 워터마크 추가하지 않기",
"Do not do anything special": "아무것도 하지 않기",
"Do not save grids consisting of one picture": "이미지가 1개뿐인 그리드는 저장하지 않기",
@ -259,6 +266,7 @@
"Keep -1 for seeds": "시드값 -1로 유지",
"keep whatever was there originally": "이미지 원본 유지",
"keyword": "프롬프트",
"Krita Plugin.": "Kirta 플러그인입니다.",
"Label": "라벨",
"Lanczos": "Lanczos",
"Last prompt:": "마지막 프롬프트 : ",
@ -303,7 +311,7 @@
"Modules": "모듈",
"Move face restoration model from VRAM into RAM after processing": "처리가 완료되면 얼굴 보정 모델을 VRAM에서 RAM으로 옮기기",
"Move to favorites": "즐겨찾기로 옮기기",
"Move VAE and CLIP to RAM when training hypernetwork. Saves VRAM.": "하이퍼네트워크 훈련 진행 시 VAE와 CLIP을 RAM으로 옮기기. VRAM이 절약됩니다.",
"Move VAE and CLIP to RAM when training if possible. Saves VRAM.": "훈련 진행 시 가능하면 VAE와 CLIP을 RAM으로 옮기기. VRAM이 절약됩니다.",
"Moved to favorites": "즐겨찾기로 옮겨짐",
"Multiplier (M) - set to 0 to get model A": "배율 (M) - 0으로 적용하면 모델 A를 얻게 됩니다",
"Name": "이름",
@ -314,6 +322,7 @@
"None": "없음",
"Nothing": "없음",
"Nothing found in the image.": "Nothing found in the image.",
"novelai-2-local-prompt": "NovelAI 프롬프트 변환기",
"Number of columns on the page": "각 페이지마다 표시할 가로줄 수",
"Number of grids in each row": "각 세로줄마다 표시될 그리드 수",
"number of images to delete consecutively next": "연속적으로 삭제할 이미지 수",
@ -418,6 +427,7 @@
"Save": "저장",
"Save a copy of embedding to log directory every N steps, 0 to disable": "N스텝마다 로그 경로에 임베딩을 저장합니다, 비활성화하려면 0으로 설정하십시오.",
"Save a copy of image before applying color correction to img2img results": "이미지→이미지 결과물에 색상 보정을 진행하기 전 이미지의 복사본을 저장하기",
"Save a copy of image before applying highres fix.": "고해상도 보정을 진행하기 전 이미지의 복사본을 저장하기",
"Save a copy of image before doing face restoration.": "얼굴 보정을 진행하기 전 이미지의 복사본을 저장하기",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "N스텝마다 로그 경로에 손실(Loss)을 포함하는 csv 파일을 저장합니다, 비활성화하려면 0으로 설정하십시오.",
"Save an image to log directory every N steps, 0 to disable": "N스텝마다 로그 경로에 이미지를 저장합니다, 비활성화하려면 0으로 설정하십시오.",
@ -427,6 +437,7 @@
"Save images with embedding in PNG chunks": "PNG 청크로 이미지에 임베딩을 포함시켜 저장",
"Save style": "스타일 저장",
"Save text information about generation parameters as chunks to png files": "이미지 생성 설정값을 PNG 청크에 텍스트로 저장",
"Saves Optimizer state as separate *.optim file. Training can be resumed with HN itself and matching optim file.": "옵티마이저 상태를 별개의 *.optim 파일로 저장하기. 하이퍼네트워크 파일과 일치하는 optim 파일로부터 훈련을 재개할 수 있습니다.",
"Saving images/grids": "이미지/그리드 저장",
"Saving to a directory": "디렉토리에 저장",
"Scale by": "스케일링 배수 지정",
@ -439,8 +450,8 @@
"See": "자세한 설명은",
"Seed": "시드",
"Seed of a different picture to be mixed into the generation.": "결과물에 섞일 다른 그림의 시드",
"Select activation function of hypernetwork": "하이퍼네트워크 활성화 함수 선택",
"Select Layer weights initialization. relu-like - Kaiming, sigmoid-like - Xavier is recommended": "레이어 가중치 초기화 방식 선택 - relu류 : Kaiming 추천, sigmoid류 : Xavier 추천",
"Select activation function of hypernetwork. Recommended : Swish / Linear(none)": "하이퍼네트워크 활성화 함수 선택 - 추천 : Swish / Linear(None)",
"Select Layer weights initialization. Recommended: Kaiming for relu-like, Xavier for sigmoid-like, Normal otherwise": "레이어 가중치 초기화 방식 선택 - relu류 : Kaiming 추천, sigmoid류 : Xavier 추천, 그 외 : Normal",
"Select which Real-ESRGAN models to show in the web UI. (Requires restart)": "WebUI에 표시할 Real-ESRGAN 모델을 선택하십시오. (재시작 필요)",
"Send seed when sending prompt or image to other interface": "다른 화면으로 프롬프트나 이미지를 보낼 때 시드도 함께 보내기",
"Send to extras": "부가기능으로 전송",
@ -456,7 +467,7 @@
"should be 2 or lower.": "이 2 이하여야 합니다.",
"Show generation progress in window title.": "창 타이틀에 생성 진행도 보여주기",
"Show grid in results for web": "웹에서 결과창에 그리드 보여주기",
"Show image creation progress every N sampling steps. Set 0 to disable.": "N번째 샘플링 스텝마다 이미지 생성 과정 보이기 - 비활성화하려면 0으로 설정",
"Show image creation progress every N sampling steps. Set to 0 to disable. Set to -1 to show after completion of batch.": "N번째 샘플링 스텝마다 이미지 생성 과정 보이기 - 비활성화하려면 0으로 설정, 배치 생성 완료시 보이려면 -1로 설정",
"Show images zoomed in by default in full page image viewer": "전체 페이지 이미지 뷰어에서 기본값으로 이미지 확대해서 보여주기",
"Show previews of all images generated in a batch as a grid": "배치에서 생성된 모든 이미지의 미리보기를 그리드 형식으로 보여주기",
"Show progressbar": "프로그레스 바 보이기",
@ -511,6 +522,7 @@
"Tile size for ESRGAN upscalers. 0 = no tiling.": "ESRGAN 업스케일러들의 타일 사이즈. 0 = 타일링 없음.",
"Tiling": "타일링",
"Time taken:": "소요 시간 : ",
"tokenizer": "토크나이저",
"Torch active/reserved:": "활성화/예약된 Torch 양 : ",
"Torch active: Peak amount of VRAM used by Torch during generation, excluding cached data.\nTorch reserved: Peak amount of VRAM allocated by Torch, including all active and cached data.\nSys VRAM: Peak amount of VRAM allocation across all applications / total GPU VRAM (peak utilization%).": "활성화된 Torch : 생성 도중 캐시된 데이터를 포함해 사용된 VRAM의 최대량\n예약된 Torch : 활성화되고 캐시된 모든 데이터를 포함해 Torch에게 할당된 VRAM의 최대량\n시스템 VRAM : 모든 어플리케이션에 할당된 VRAM 최대량 / 총 GPU VRAM (최고 이용도%)",
"Train": "훈련",
@ -518,6 +530,7 @@
"Train Embedding": "임베딩 훈련",
"Train Hypernetwork": "하이퍼네트워크 훈련",
"Training": "훈련",
"training-picker": "훈련용 선택기",
"txt2img": "텍스트→이미지",
"txt2img history": "텍스트→이미지 기록",
"uniform": "uniform",
@ -541,12 +554,14 @@
"Use BLIP for caption": "캡션에 BLIP 사용",
"Use checkbox to enable the extension; it will be enabled or disabled when you click apply button": "체크박스를 이용해 적용할 확장기능을 선택하세요. 변경사항은 적용 후 UI 재시작 버튼을 눌러야 적용됩니다.",
"Use checkbox to mark the extension for update; it will be updated when you click apply button": "체크박스를 이용해 업데이트할 확장기능을 선택하세요. 업데이트는 적용 후 UI 재시작 버튼을 눌러야 적용됩니다.",
"Use cross attention optimizations while training": "훈련 진행 시 크로스 어텐션 최적화 사용",
"Use deepbooru for caption": "캡션에 deepbooru 사용",
"Use dropout": "드롭아웃 사용",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "다음 태그들을 사용해 이미지 파일명 형식을 결정하세요 : [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]. 비워두면 기본값으로 설정됩니다.",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "다음 태그들을 사용해 이미지와 그리드의 하위 디렉토리명의 형식을 결정하세요 : [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]. 비워두면 기본값으로 설정됩니다.",
"Use old emphasis implementation. Can be useful to reproduce old seeds.": "옛 방식의 강조 구현을 사용합니다. 옛 시드를 재현하는 데 효과적일 수 있습니다.",
"Use original name for output filename during batch process in extras tab": "부가기능 탭에서 이미지를 여러장 처리 시 결과물 파일명에 기존 파일명 사용하기",
"Use same random seed for all lines": "모든 줄에 동일한 시드 사용",
"Use same seed for each image": "각 이미지에 동일한 시드 사용",
"use spaces for tags in deepbooru": "deepbooru에서 태그에 공백 사용",
"User interface": "사용자 인터페이스",

View file

@ -17,6 +17,7 @@
"Checkpoint Merger": "Fusão de Checkpoint",
"Train": "Treinar",
"Settings": "Configurações",
"Extensions": "Extensões",
"Prompt": "Prompt",
"Negative prompt": "Prompt negativo",
"Run": "Executar",
@ -56,7 +57,7 @@
"Highres. fix": "Ajuste de alta resolução",
"Firstpass width": "Primeira Passagem da largura",
"Firstpass height": "Primeira Passagem da altura",
"Denoising strength": "Denoising strength",
"Denoising strength": "Força do denoise",
"Batch count": "Quantidade por lote",
"Batch size": "Quantidade de lotes",
"CFG Scale": "Escala CFG",
@ -93,13 +94,13 @@
"Eta": "Tempo estimado",
"Clip skip": "Pular Clip",
"Denoising": "Denoising",
"Cond. Image Mask Weight": "Peso da Máscara Condicional de Imagem",
"X values": "Valores de X",
"Y type": "Tipo de Y",
"Y values": "Valores de Y",
"Draw legend": "Desenhar a legenda",
"Include Separate Images": "Incluir Imagens Separadas",
"Keep -1 for seeds": "Manter em -1 para seeds",
"Drop Image Here": "Solte a imagem aqui",
"Save": "Salvar",
"Send to img2img": "Mandar para img2img",
"Send to inpaint": "Mandar para inpaint",
@ -110,6 +111,7 @@
"Inpaint": "Inpaint",
"Batch img2img": "Lote img2img",
"Image for img2img": "Imagem para img2img",
"Drop Image Here": "Solte a imagem aqui",
"Image for inpainting with mask": "Imagem para inpainting com máscara",
"Mask": "Máscara",
"Mask blur": "Desfoque da máscara",
@ -166,16 +168,10 @@
"Upscaler": "Ampliador",
"Lanczos": "Lanczos",
"LDSR": "LDSR",
"4x_foolhardy_Remacri": "4x_foolhardy_Remacri",
"Put ESRGAN models here": "Coloque modelos ESRGAN aqui",
"R-ESRGAN General 4xV3": "R-ESRGAN General 4xV3",
"R-ESRGAN AnimeVideo": "R-ESRGAN AnimeVideo",
"R-ESRGAN 4x+": "R-ESRGAN 4x+",
"R-ESRGAN 4x+ Anime6B": "R-ESRGAN 4x+ Anime6B",
"R-ESRGAN 2x+": "R-ESRGAN 2x+",
"ScuNET": "ScuNET",
"ESRGAN_4x": "ESRGAN_4x",
"ScuNET GAN": "ScuNET GAN",
"ScuNET PSNR": "ScuNET PSNR",
"put_swinir_models_here": "put_swinir_models_here",
"SwinIR 4x": "SwinIR 4x",
"Single Image": "Uma imagem",
"Batch Process": "Processo em lote",
"Batch from Directory": "Lote apartir de diretório",
@ -189,7 +185,7 @@
"GFPGAN visibility": "Visibilidade GFPGAN",
"CodeFormer visibility": "Visibilidade CodeFormer",
"CodeFormer weight (0 = maximum effect, 1 = minimum effect)": "Peso do CodeFormer (0 = efeito máximo, 1 = efeito mínimo)",
"Open output directory": "Abrir diretório de saída",
"Upscale Before Restoring Faces": "Ampliar Antes de Refinar Rostos",
"Send to txt2img": "Mandar para txt2img",
"A merger of the two checkpoints will be generated in your": "Uma fusão dos dois checkpoints será gerada em seu",
"checkpoint": "checkpoint",
@ -216,6 +212,7 @@
"Modules": "Módulos",
"Enter hypernetwork layer structure": "Entrar na estrutura de camadas da hypernetwork",
"Select activation function of hypernetwork": "Selecionar a função de ativação de hypernetwork",
"linear": "linear",
"relu": "relu",
"leakyrelu": "leakyrelu",
"elu": "elu",
@ -227,12 +224,10 @@
"glu": "glu",
"hardshrink": "hardshrink",
"hardsigmoid": "hardsigmoid",
"hardswish": "hardswish",
"hardtanh": "hardtanh",
"logsigmoid": "logsigmoid",
"logsoftmax": "logsoftmax",
"mish": "mish",
"multiheadattention": "multiheadattention",
"prelu": "prelu",
"rrelu": "rrelu",
"relu6": "relu6",
@ -274,9 +269,9 @@
"Focal point edges weight": "Peso de ponto focal para bordas",
"Create debug image": "Criar imagem de depuração",
"Preprocess": "Pré-processar",
"Train an embedding; must specify a directory with a set of 1:1 ratio images": "Treinar um embedding; precisa especificar um diretório com imagens de proporção 1:1",
"Train an embedding; must specify a directory with a set of 1:1 ratio images": "Treinar uma incorporação; precisa especificar um diretório com imagens de proporção 1:1",
"[wiki]": "[wiki]",
"Embedding": "Embedding",
"Embedding": "Incorporação",
"Embedding Learning rate": "Taxa de aprendizagem da incorporação",
"Hypernetwork Learning rate": "Taxa de aprendizagem de Hypernetwork",
"Dataset directory": "Diretório de Dataset",
@ -345,9 +340,11 @@
"Filename join string": "Nome de arquivo join string",
"Number of repeats for a single input image per epoch; used only for displaying epoch number": "Número de repetições para entrada única de imagens por época; serve apenas para mostrar o número de época",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "Salvar um csv com as perdas para o diretório de log a cada N steps, 0 para desativar",
"Use cross attention optimizations while training": "Usar otimizações de atenção cruzada enquanto treinando",
"Stable Diffusion": "Stable Diffusion",
"Checkpoints to cache in RAM": "Checkpoints para manter no cache da RAM",
"Hypernetwork strength": "Força da Hypernetwork",
"Inpainting conditioning mask strength": "Força do inpaint para máscaras condicioniais",
"Apply color correction to img2img results to match original colors.": "Aplicar correção de cor nas imagens geradas em img2img, usando a imagem original como base.",
"Save a copy of image before applying color correction to img2img results": "Salvar uma cópia das imagens geradas em img2img antes de aplicar a correção de cor",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "Durante gerações img2img, fazer examente o número de steps definidos na barra (normalmente você faz menos steps com denoising menor).",
@ -379,6 +376,7 @@
"Add model hash to generation information": "Adicionar hash do modelo para informação de geração",
"Add model name to generation information": "Adicionar nome do modelo para informação de geração",
"When reading generation parameters from text into UI (from PNG info or pasted text), do not change the selected model/checkpoint.": "Quando ler parâmetros de texto para a interface (de informações de PNG ou texto copiado), não alterar o modelo/intervalo selecionado.",
"Send seed when sending prompt or image to other interface": "Enviar seed quando enviar prompt ou imagem para outra interface",
"Font for image grids that have text": "Fonte para grade de imagens que têm texto",
"Enable full page image viewer": "Ativar visualizador de página inteira",
"Show images zoomed in by default in full page image viewer": "Mostrar imagens com zoom por definição no visualizador de página inteira",
@ -386,13 +384,17 @@
"Quicksettings list": "Lista de configurações rapidas",
"Localization (requires restart)": "Localização (precisa reiniciar)",
"ar_AR": "ar_AR",
"de_DE": "de_DE",
"es_ES": "es_ES",
"fr-FR": "fr-FR",
"fr_FR": "fr_FR",
"it_IT": "it_IT",
"ja_JP": "ja_JP",
"ko_KR": "ko_KR",
"pt_BR": "pt_BR",
"ru_RU": "ru_RU",
"tr_TR": "tr_TR",
"zh_CN": "zh_CN",
"zh_TW": "zh_TW",
"Sampler parameters": "Parâmetros de Amostragem",
"Hide samplers in user interface (requires restart)": "Esconder amostragens na interface de usuário (precisa reiniciar)",
"eta (noise multiplier) for DDIM": "tempo estimado (multiplicador de ruído) para DDIM",
@ -408,6 +410,19 @@
"Download localization template": "Baixar arquivo modelo de localização",
"Reload custom script bodies (No ui updates, No restart)": "Recarregar scripts personalizados (Sem atualizar a interface, Sem reiniciar)",
"Restart Gradio and Refresh components (Custom Scripts, ui.py, js and css only)": "Reiniciar Gradio e atualizar componentes (Scripts personalizados, ui.py, js e css)",
"Installed": "Instalado",
"Available": "Disponível",
"Install from URL": "Instalado de URL",
"Apply and restart UI": "Apicar e reiniciar a interface",
"Check for updates": "Procurar por atualizações",
"Extension": "Extensão",
"URL": "URL",
"Update": "Atualização",
"Load from:": "Carregar de:",
"Extension index URL": "Índice de extensão URL",
"URL for extension's git repository": "URL para repositório git da extensão",
"Local directory name": "Nome do diretório local",
"Install": "Instalar",
"Prompt (press Ctrl+Enter or Alt+Enter to generate)": "Prompt (apertar Ctrl+Enter ou Alt+Enter para gerar)",
"Negative prompt (press Ctrl+Enter or Alt+Enter to generate)": "Prompt Negativo (apertar Ctrl+Enter ou Alt+Enter para gerar)",
"Add a random artist to the prompt.": "Adicionar um artista aleatório para o prompt.",
@ -420,7 +435,7 @@
"Do not do anything special": "Não faça nada de especial",
"Which algorithm to use to produce the image": "O tipo de algoritmo para gerar imagens.",
"Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help": "Euler Ancestral - cria mais variações para as imagens em diferentes passos. Mais que 40 passos cancela o efeito.",
"Denoising Diffusion Implicit Models - Funciona melhor para inpainting.": "Denoising Diffusion Implicit Models - Funciona melhor para inpainting.",
"Denoising Diffusion Implicit Models - best at inpainting": "Denoising Diffusion Implicit Models - Funciona melhor para inpainting.",
"Produce an image that can be tiled.": "Produz uma imagem que pode ser ladrilhada.",
"Use a two step process to partially create an image at smaller resolution, upscale, and then improve details in it without changing composition": "Cria um processo em duas etapas, com uma imagem em baixa qualidade primeiro, aumenta a imagem e refina os detalhes sem alterar a composição da imagem",
"Determines how little respect the algorithm should have for image's content. At 0, nothing will change, and at 1 you'll get an unrelated image. With values below 1.0, processing will take less steps than the Sampling Steps slider specifies.": "Quanto o algoritmo deve manter da imagem original. Em 0, nada muda. Em 1 o algoritmo ignora a imagem original. Valores menores que 1.0 demoram mais.",
@ -438,7 +453,7 @@
"Write image to a directory (default - log/images) and generation parameters into csv file.": "Salva a imagem no diretório padrão ou escolhido e cria um arquivo csv com os parâmetros da geração.",
"Open images output directory": "Abre o diretório de saída de imagens.",
"How much to blur the mask before processing, in pixels.": "Transição do contorno da máscara, em pixels.",
"What to put inside the masked area before processing it with Stable Diffusion.": "O que vai dentro da máscara antes de processar.",
"What to put inside the masked area before processing it with Stable Diffusion.": "O que vai dentro da máscara antes de processá-la com Stable Diffusion.",
"fill it with colors of the image": "Preenche usando as cores da imagem.",
"keep whatever was there originally": "manter usando o que estava lá originalmente",
"fill it with latent space noise": "Preenche com ruídos do espaço latente.",
@ -463,6 +478,8 @@
"Restore low quality faces using GFPGAN neural network": "Restaurar rostos de baixa qualidade usando a rede neural GFPGAN",
"This regular expression will be used extract words from filename, and they will be joined using the option below into label text used for training. Leave empty to keep filename text as it is.": "Esta expressão regular vai retirar palavras do nome do arquivo e serão juntadas via regex usando a opção abaixo em etiquetas usadas em treinamento. Não mexer para manter os nomes como estão.",
"This string will be used to join split words into a single line if the option above is enabled.": "Esta string será usada para unir palavras divididas em uma única linha se a opção acima estiver habilitada.",
"Only applies to inpainting models. Determines how strongly to mask off the original image for inpainting and img2img. 1.0 means fully masked, which is the default behaviour. 0.0 means a fully unmasked conditioning. Lower values will help preserve the overall composition of the image, but will struggle with large changes.": "Aplicável somente para modelos de inpaint. Determina quanto deve mascarar da imagem original para inpaint e img2img. 1.0 significa totalmente mascarado, que é o comportamento padrão. 0.0 significa uma condição totalmente não mascarada. Valores baixos ajudam a preservar a composição geral da imagem, mas vai encontrar dificuldades com grandes mudanças.",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "Lista de nomes de configurações, separados por vírgulas, para configurações que devem ir para a barra de acesso rápido na parte superior, em vez da guia de configuração usual. Veja modules/shared.py para nomes de configuração. Necessita reinicialização para aplicar.",
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "Se este valor for diferente de zero, ele será adicionado à seed e usado para inicializar o RNG para ruídos ao usar amostragens com Tempo Estimado. Você pode usar isso para produzir ainda mais variações de imagens ou pode usar isso para combinar imagens de outro software se souber o que está fazendo."
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "Se este valor for diferente de zero, ele será adicionado à seed e usado para inicializar o RNG para ruídos ao usar amostragens com Tempo Estimado. Você pode usar isso para produzir ainda mais variações de imagens ou pode usar isso para combinar imagens de outro software se souber o que está fazendo.",
"Leave empty for auto": "Deixar desmarcado para automático"
}

View file

@ -7,19 +7,19 @@
"Loading...": "载入中...",
"view": "查看",
"api": "api",
"•": "",
"•": "",
"built with gradio": "基于 Gradio 构建",
"Stable Diffusion checkpoint": "Stable Diffusion 模型(ckpt)",
"txt2img": "文生图",
"img2img": "图生图",
"Extras": "后处理",
"PNG Info": "PNG 信息",
"Checkpoint Merger": "模型(ckpt)合并工具",
"Extras": "更多",
"PNG Info": "图片信息",
"Checkpoint Merger": "模型(ckpt)合并",
"Train": "训练",
"Create aesthetic embedding": "生成美术风格 embedding",
"Create aesthetic embedding": "生成美术风格",
"Image Browser": "图库浏览器",
"History": "历史记录",
"Settings": "设置",
"Extensions": "扩展",
"Prompt": "提示词",
"Negative prompt": "反向提示词",
"Run": "运行",
@ -37,8 +37,8 @@
"Image": "图像",
"Check progress": "查看进度",
"Check progress (first)": "(首次)查看进度",
"Sampling Steps": "采样迭代步数",
"Sampling method": "采样方法",
"Sampling Steps": "采样迭代步数 (Steps)",
"Sampling method": "采样方法 (Sampler)",
"Euler a": "Euler a",
"Euler": "Euler",
"LMS": "LMS",
@ -59,38 +59,39 @@
"Highres. fix": "高分辨率修复",
"Firstpass width": "第一遍的宽度",
"Firstpass height": "第一遍的高度",
"Denoising strength": "去噪强度",
"Batch count": "批次",
"Batch size": "量",
"Denoising strength": "重绘幅度(Denoising strength)",
"Batch count": "生成批次",
"Batch size": "每批数量",
"CFG Scale": "提示词相关性(CFG Scale)",
"Seed": "随机种子",
"Extra": "额外参数",
"Seed": "随机种子(seed)",
"Extra": "",
"Variation seed": "差异随机种子",
"Variation strength": "差异强度",
"Resize seed from width": "自宽度缩放随机种子",
"Resize seed from height": "自高度缩放随机种子",
"Open for Clip Aesthetic!": "打开美术风格 Clip!",
"▼": "▼",
"Open for Clip Aesthetic!": "打开以调整 Clip 的美术风格!",
"Aesthetic weight": "美术风格权重",
"Aesthetic steps": "美术风格迭代步数",
"Aesthetic learning rate": "美术风格学习率",
"Slerp interpolation": "Slerp 插值",
"Slerp interpolation": "球面线性插值",
"Aesthetic imgs embedding": "美术风格图集 embedding",
"None": "无",
"Aesthetic text for imgs": "该图集的美术风格描述",
"Slerp angle": "Slerp 角度",
"Slerp angle": "球面线性插值角度",
"Is negative text": "是反向提示词",
"Script": "脚本",
"Embedding to Shareable PNG": "将 Embedding 转换为可分享的 PNG",
"Embedding to Shareable PNG": "将 Embedding 转换为可分享的 PNG 图片文件",
"Prompt matrix": "提示词矩阵",
"Prompts from file or textbox": "从文本框或文件载入提示词",
"X/Y plot": "X/Y 图表",
"Source embedding to convert": "用于转换的源 Embedding",
"Embedding token": "Embedding 的 token (关键词)",
"Output directory": "输出目录",
"Put variable parts at start of prompt": "把变量部分放在提示词文本的开头",
"Show Textbox": "显示文本框",
"File with inputs": "含输入内容的文件",
"Prompts": "提示词",
"Iterate seed every line": "每行输入都换一个种子",
"Use same random seed for all lines": "每行输入都使用同一个随机种子",
"List of prompt inputs": "提示词输入列表",
"Upload prompt inputs": "上传提示词输入文件",
"X type": "X轴类型",
"Nothing": "无",
"Var. seed": "差异随机种子",
@ -100,62 +101,61 @@
"Prompt order": "提示词顺序",
"Sampler": "采样器",
"Checkpoint name": "模型(ckpt)名",
"Hypernetwork": "Hypernetwork",
"Hypernet str.": "Hypernetwork 强度",
"Hypernetwork": "超网络(Hypernetwork)",
"Hypernet str.": "超网络(Hypernetwork) 强度",
"Sigma Churn": "Sigma Churn",
"Sigma min": "最小 Sigma",
"Sigma max": "最大 Sigma",
"Sigma noise": "Sigma noise",
"Eta": "Eta",
"Clip skip": "Clip 跳过",
"Denoising": "去噪",
"Denoising": "重绘幅度",
"Cond. Image Mask Weight": "图像调节屏蔽度",
"X values": "X轴数值",
"Y type": "Y轴类型",
"Y values": "Y轴数值",
"Draw legend": "在图表中包括轴标题",
"Include Separate Images": "包括独立的图像",
"Keep -1 for seeds": "保持随机种子为-1",
"Drop Image Here": "拖拽图像到此",
"Save": "保存",
"Send to img2img": ">> 图生图",
"Send to inpaint": ">> 内补绘制",
"Send to extras": ">> 后处理",
"Send to inpaint": ">> 局部重绘",
"Send to extras": ">> 更多",
"Make Zip when Save?": "保存时生成zip压缩文件?",
"Textbox": "文本框",
"Interrogate\nCLIP": "CLIP\n反推提示词",
"Interrogate\nDeepBooru": "DeepBooru\n反推提示词",
"Inpaint": "内补绘制",
"Inpaint": "局部重绘",
"Batch img2img": "批量图生图",
"Image for img2img": "图生图的图像",
"Image for inpainting with mask": "用于内补绘制蒙版内容的图像",
"Drop Image Here": "拖拽图像到此",
"Image for inpainting with mask": "用于局部重绘并手动画蒙版的图像",
"Mask": "蒙版",
"Mask blur": "蒙版模糊",
"Mask mode": "蒙版模式",
"Draw mask": "绘制蒙版",
"Upload mask": "上传蒙版",
"Masking mode": "蒙版模式",
"Inpaint masked": "内补绘制蒙版内容",
"Inpaint not masked": "内补绘制非蒙版内容",
"Inpaint masked": "重绘蒙版内容",
"Inpaint not masked": "重绘非蒙版内容",
"Masked content": "蒙版蒙住的内容",
"fill": "填充",
"original": "原图",
"latent noise": "潜空间噪声",
"latent nothing": "潜空间数值零",
"Inpaint at full resolution": "以完整分辨率进行内补绘制",
"Inpaint at full resolution padding, pixels": "以完整分辨率进行内补绘制 - 填补像素",
"Process images in a directory on the same machine where the server is running.": "在服务器主机上的目录中处理图像",
"Use an empty output directory to save pictures normally instead of writing to the output directory.": "指定一个空的文件夹为输出目录而非默认的 output 文件夹为输出目录",
"Disabled when launched with --hide-ui-dir-config.": "启动 --hide-ui-dir-config 时禁用",
"Inpaint at full resolution": "全分辨率局部重绘",
"Inpaint at full resolution padding, pixels": "预留像素",
"Process images in a directory on the same machine where the server is running.": "使用服务器主机上的一个目录,作为输入目录处理图像",
"Use an empty output directory to save pictures normally instead of writing to the output directory.": "使用一个空的文件夹作为输出目录,而不是使用默认的 output 文件夹作为输出目录",
"Input directory": "输入目录",
"Output directory": "输出目录",
"Resize mode": "缩放模式",
"Just resize": "只缩放",
"Crop and resize": "缩放并剪裁",
"Resize and fill": "缩放并填充",
"Just resize": "拉伸",
"Crop and resize": "",
"Resize and fill": "填充",
"img2img alternative test": "图生图的另一种测试",
"Loopback": "回送",
"Outpainting mk2": "绘制第二版",
"Poor man's outpainting": "效果稍差的绘制",
"Outpainting mk2": "外绘制第二版",
"Poor man's outpainting": "效果稍差的外绘制",
"SD upscale": "使用 SD 放大(SD upscale)",
"should be 2 or lower.": "必须小于等于2",
"Override `Sampling method` to Euler?(this method is built for it)": "覆写 `采样方法` 为 Euler?(这个方法就是为这样做设计的)",
@ -164,15 +164,15 @@
"Original negative prompt": "初始反向提示词",
"Override `Sampling Steps` to the same value as `Decode steps`?": "覆写 `采样迭代步数` 为 `解码迭代步数`?",
"Decode steps": "解码迭代步数",
"Override `Denoising strength` to 1?": "覆写 `去噪强度` 为 1?",
"Override `Denoising strength` to 1?": "覆写 `重绘幅度` 为 1?",
"Decode CFG scale": "解码提示词相关性(CFG scale)",
"Randomness": "随机度",
"Sigma adjustment for finding noise for image": "为寻找图中噪点的 Sigma 调整",
"Loops": "迭代次数",
"Denoising strength change factor": "去噪强度的调整系数",
"Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8": "推荐设置采样迭代步数80-100采样器Euler a去噪强0.8",
"Denoising strength change factor": "重绘幅度的调整系数",
"Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8": "推荐设置采样迭代步数80-100采样器Euler a重绘幅0.8",
"Pixels to expand": "拓展的像素数",
"Outpainting direction": "绘制的方向",
"Outpainting direction": "外绘制的方向",
"left": "左",
"right": "右",
"up": "上",
@ -183,6 +183,7 @@
"Tile overlap": "图块重叠的像素(Tile overlap)",
"Upscaler": "放大算法",
"Lanczos": "Lanczos",
"Nearest": "最邻近(整数缩放)",
"LDSR": "LDSR",
"BSRGAN 4x": "BSRGAN 4x",
"ESRGAN_4x": "ESRGAN_4x",
@ -199,12 +200,11 @@
"Scale to": "指定尺寸缩放",
"Resize": "缩放",
"Crop to fit": "裁剪以适应",
"Upscaler 2": "放大算法 2",
"Upscaler 2 visibility": "放大算法 2 可见度",
"Upscaler 2 visibility": "放大算法 2 (Upscaler 2) 可见度",
"GFPGAN visibility": "GFPGAN 可见度",
"CodeFormer visibility": "CodeFormer 可见度",
"CodeFormer weight (0 = maximum effect, 1 = minimum effect)": "CodeFormer 权重 (0 = 最大效果, 1 = 最小效果)",
"Open output directory": "打开输出目录",
"Upscale Before Restoring Faces": "放大后再进行面部修复",
"Send to txt2img": ">> 文生图",
"A merger of the two checkpoints will be generated in your": "合并后的模型(ckpt)会生成在你的",
"checkpoint": "模型(ckpt)",
@ -219,10 +219,9 @@
"Add difference": "添加差分",
"Save as float16": "以 float16 储存",
"See": "查看",
"wiki": "wiki",
"wiki": "wiki文档",
"for detailed explanation.": "以了解详细说明",
"Create embedding": "生成 embedding",
"Create aesthetic images embedding": "生成美术风格图集 embedding",
"Create hypernetwork": "生成 hypernetwork",
"Preprocess images": "图像预处理",
"Name": "名称",
@ -237,6 +236,36 @@
"leakyrelu": "leakyrelu",
"elu": "elu",
"swish": "swish",
"tanh": "tanh",
"sigmoid": "sigmoid",
"celu": "celu",
"gelu": "gelu",
"glu": "glu",
"hardshrink": "hardshrink",
"hardsigmoid": "hardsigmoid",
"hardtanh": "hardtanh",
"logsigmoid": "logsigmoid",
"logsoftmax": "logsoftmax",
"mish": "mish",
"prelu": "prelu",
"rrelu": "rrelu",
"relu6": "relu6",
"selu": "selu",
"silu": "silu",
"softmax": "softmax",
"softmax2d": "softmax2d",
"softmin": "softmin",
"softplus": "softplus",
"softshrink": "softshrink",
"softsign": "softsign",
"tanhshrink": "tanhshrink",
"threshold": "阈值",
"Select Layer weights initialization. relu-like - Kaiming, sigmoid-like - Xavier is recommended": "选择初始化层权重的方案. 类relu - Kaiming, 类sigmoid - Xavier 都是比较推荐的选项",
"Normal": "正态",
"KaimingUniform": "Kaiming 均匀",
"KaimingNormal": "Kaiming 正态",
"XavierUniform": "Xavier 均匀",
"XavierNormal": "Xavier 正态",
"Add layer normalization": "添加层标准化",
"Use dropout": "采用 dropout 防止过拟合",
"Overwrite Old Hypernetwork": "覆写旧的 Hypernetwork",
@ -248,89 +277,92 @@
"prepend": "放前面",
"append": "放后面",
"Create flipped copies": "生成镜像副本",
"Split oversized images into two": "将过大的图像分为两份",
"Split oversized images": "分割过大的图像",
"Auto focal point crop": "自动焦点裁切",
"Use BLIP for caption": "使用 BLIP 生成说明文字(自然语言描述)",
"Use deepbooru for caption": "使用 deepbooru 生成说明文字(tags)",
"Split image threshold": "图像分割阈值",
"Split image overlap ratio": "分割图像重叠的比率",
"Focal point face weight": "焦点面部权重",
"Focal point entropy weight": "焦点熵权重",
"Focal point edges weight": "焦点线条权重",
"Create debug image": "生成调试(debug)图片",
"Preprocess": "预处理",
"Train an embedding; must specify a directory with a set of 1:1 ratio images": "训练 embedding 必须指定一组具有 1:1 比例图像的目录",
"Train an embedding or Hypernetwork; you must specify a directory with a set of 1:1 ratio images": "训练 embedding 或者 hypernetwork 必须指定一组具有 1:1 比例图像的目录",
"[wiki]": "[wiki]",
"[wiki]": "[wiki文档]",
"Embedding": "Embedding",
"Embedding Learning rate": "Embedding 学习率",
"Hypernetwork Learning rate": "Hypernetwork 学习率",
"Learning rate": "学习率",
"Dataset directory": "数据集目录",
"Log directory": "日志目录",
"Prompt template file": "提示词模版文件",
"Max steps": "最大迭代步数",
"Save an image to log directory every N steps, 0 to disable": "每 N 步保存一个图像到日志目录0 表示禁用",
"Save a copy of embedding to log directory every N steps, 0 to disable": "每 N 步将 embedding 的副本保存到日志目录0 表示禁用",
"Save images with embedding in PNG chunks": "保存图像并在 PNG 文件中嵌入 embedding 文件",
"Read parameters (prompt, etc...) from txt2img tab when making previews": "进行预览时从文生图选项卡中读取参数(提示词等)",
"Save images with embedding in PNG chunks": "保存图像并在 PNG 图片文件中嵌入 embedding 文件",
"Read parameters (prompt, etc...) from txt2img tab when making previews": "进行预览时从文生图选项卡中读取参数(提示词等)",
"Train Hypernetwork": "训练 Hypernetwork",
"Train Embedding": "训练 Embedding",
"Create an aesthetic embedding out of any number of images": "从任意数量的图像中创建美术风格 embedding",
"Create images embedding": "生成图集 embedding",
"txt2img history": "文生图历史记录",
"img2img history": "图生图历史记录",
"extras history": "后处理历史记录",
"Renew Page": "刷新页面",
"extras": "后处理",
"favorites": "收藏夹",
"custom fold": "自定义文件夹",
"Load": "载入",
"Favorites": "收藏夹(已保存)",
"Others": "其他",
"Images directory": "图像目录",
"Prev batch": "上一批",
"Next batch": "下一批",
"Dropdown": "下拉列表",
"First Page": "首页",
"Prev Page": "上一页",
"Page Index": "页数",
"Next Page": "下一页",
"End Page": "尾页",
"number of images to delete consecutively next": "接下来要连续删除的图像数",
"delete next": "删除下一张",
"Delete": "删除",
"sort by": "排序方式",
"path name": "路径名",
"date": "日期",
"keyword": "搜索",
"Generate Info": "生成信息",
"File Name": "文件名",
"Collect": "收藏",
"Refresh page": "刷新页面",
"Date to": "日期至",
"Move to favorites": "移动到收藏夹(保存)",
"Renew Page": "刷新页面",
"Number": "数量",
"set_index": "设置索引",
"load_switch": "载入开关",
"turn_page_switch": "翻页开关",
"Checkbox": "勾选框",
"Apply settings": "保存设置",
"Saving images/grids": "保存图像/概览图",
"Saving images/grids": "保存图像/宫格图",
"Always save all generated images": "始终保存所有生成的图像",
"File format for images": "图像的文件格式",
"Images filename pattern": "图像文件名格式",
"Always save all generated image grids": "始终保存所有生成的概览图",
"File format for grids": "概览图的文件格式",
"Add extended info (seed, prompt) to filename when saving grid": "保存概览时将扩展信息(随机种子、提示词)添加到文件名",
"Do not save grids consisting of one picture": "只有一张图片时不要保存概览图",
"Prevent empty spots in grid (when set to autodetect)": "(在自动检测时)防止概览图中出现空位",
"Grid row count; use -1 for autodetect and 0 for it to be same as batch size": "概览行数; 使用 -1 进行自动检测,使用 0 使其与批量大小相同",
"Save text information about generation parameters as chunks to png files": "将有关生成参数的文本信息作为块保存到 png 文件中",
"Create a text file next to every image with generation parameters.": "保存图像时在每个图像旁边创建一个文本文件储存生成参数",
"Add number to filename when saving": "储存的时候在文件名里添加数字",
"Always save all generated image grids": "始终保存所有生成的宫格图",
"File format for grids": "宫格图的文件格式",
"Add extended info (seed, prompt) to filename when saving grid": "保存宫格图时,将扩展信息(随机种子、提示词)添加到文件名",
"Do not save grids consisting of one picture": "只有一张图片时不要保存宫格图",
"Prevent empty spots in grid (when set to autodetect)": "(启用自动检测时)防止宫格图中出现空位",
"Grid row count; use -1 for autodetect and 0 for it to be same as batch size": "宫格图行数; 使用 -1 进行自动检测,使用 0 使其与每批数量相同",
"Save text information about generation parameters as chunks to png files": "将有关生成参数的文本信息,作为块保存到 png 图片文件中",
"Create a text file next to every image with generation parameters.": "保存图像时,在每个图像旁边创建一个文本文件储存生成参数",
"Save a copy of image before doing face restoration.": "在进行面部修复之前保存图像副本",
"Save a copy of image before applying highres fix.": "在做高分辨率修复之前保存初始图像副本",
"Save a copy of image before applying color correction to img2img results": "在对图生图结果应用颜色校正之前保存图像副本",
"Quality for saved jpeg images": "保存的 jpeg 图像的质量",
"If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG": "如果 PNG 图像大于 4MB 或宽高大于 4000则缩小并保存副本为 JPG",
"Use original name for output filename during batch process in extras tab": "在后处理选项卡中的批量处理过程中使用原始名称作为输出文件名",
"If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG": "如果 PNG 图像大于 4MB 或宽高大于 4000则缩小并保存副本为 JPG 图片",
"Use original name for output filename during batch process in extras tab": "在更多选项卡中的批量处理过程中,使用原始名称作为输出文件名",
"When using 'Save' button, only save a single selected image": "使用“保存”按钮时,只保存一个选定的图像",
"Do not add watermark to images": "不要给图像加水印",
"Paths for saving": "保存路径",
"Output directory for images; if empty, defaults to three directories below": "图像的输出目录; 如果为空,则默认为以下三个目录",
"Output directory for txt2img images": "文生图的输出目录",
"Output directory for img2img images": "图生图的输出目录",
"Output directory for images from extras tab": "后处理的输出目录",
"Output directory for grids; if empty, defaults to two directories below": "概览图的输出目录; 如果为空,则默认为以下两个目录",
"Output directory for txt2img grids": "文生图概览的输出目录",
"Output directory for img2img grids": "图生图概览的输出目录",
"Output directory for images from extras tab": "更多选项卡的输出目录",
"Output directory for grids; if empty, defaults to two directories below": "宫格图的输出目录; 如果为空,则默认为以下两个目录",
"Output directory for txt2img grids": "文生图宫格的输出目录",
"Output directory for img2img grids": "图生图宫格的输出目录",
"Directory for saving images using the Save button": "使用“保存”按钮保存图像的目录",
"Saving to a directory": "保存到目录",
"Save images to a subdirectory": "将图像保存到子目录",
"Save grids to a subdirectory": "将概览图保存到子目录",
"Save grids to a subdirectory": "将宫格图保存到子目录",
"When using \"Save\" button, save images to a subdirectory": "使用“保存”按钮时,将图像保存到子目录",
"Directory name pattern": "目录名称格式",
"Max prompt words for [prompt_words] pattern": "[prompt_words] 格式的最大提示词数量",
@ -341,34 +373,36 @@
"Tile overlap, in pixels for SwinIR. Low values = visible seam.": "SwinIR 的图块重叠(Tile overlap)像素。低值 = 可见接缝",
"LDSR processing steps. Lower = faster": "LDSR 处理迭代步数。更低 = 更快",
"Upscaler for img2img": "图生图的放大算法",
"Upscale latent space image when doing hires. fix": "做高分辨率修复时也放大潜空间图像",
"Upscale latent space image when doing hires. fix": "做高分辨率修复时也放大潜空间图像",
"Face restoration": "面部修复",
"CodeFormer weight parameter; 0 = maximum effect; 1 = minimum effect": "CodeFormer 权重参数; 0 = 最大效果; 1 = 最小效果",
"Move face restoration model from VRAM into RAM after processing": "面部修复处理完成后将面部修复模型从显存(VRAM)移至内存(RAM)",
"Move face restoration model from VRAM into RAM after processing": "面部修复处理完成后将面部修复模型从显存(VRAM)移至内存(RAM)",
"System": "系统",
"VRAM usage polls per second during generation. Set to 0 to disable.": "生成图像时每秒轮询显存(VRAM)使用情况的次数。设置为 0 以禁用",
"VRAM usage polls per second during generation. Set to 0 to disable.": "生成图像时每秒轮询显存(VRAM)使用情况的次数。设置为 0 以禁用",
"Always print all generation info to standard output": "始终将所有生成信息输出到 standard output (一般为控制台)",
"Add a second progress bar to the console that shows progress for an entire job.": "向控制台添加第二个进度条,显示整个作业的进度",
"Training": "训练",
"Unload VAE and CLIP from VRAM when training": "训练时从显存(VRAM)中取消 VAE 和 CLIP 的加载",
"Move VAE and CLIP to RAM when training hypernetwork. Saves VRAM.": "训练时将 VAE 和 CLIP 从显存(VRAM)移放到内存(RAM),节省显存(VRAM)",
"Move VAE and CLIP to RAM when training if possible. Saves VRAM.": "训练时将 VAE 和 CLIP 从显存(VRAM)移放到内存(RAM)如果可行的话,节省显存(VRAM)",
"Filename word regex": "文件名用词的正则表达式",
"Filename join string": "文件名连接用字符串",
"Number of repeats for a single input image per epoch; used only for displaying epoch number": "每个 epoch 中单个输入图像的重复次数; 仅用于显示 epoch 数",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "每 N 步保存一个包含 loss 的 csv 到日志目录0 表示禁用",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "每 N 步保存一个包含 loss 的 csv 表格到日志目录0 表示禁用",
"Use cross attention optimizations while training": "训练时开启 cross attention 优化",
"Stable Diffusion": "Stable Diffusion",
"Checkpoints to cache in RAM": "缓存在内存(RAM)中的模型(ckpt)",
"SD VAE": "模型的 VAE (SD VAE)",
"auto": "自动",
"Hypernetwork strength": "Hypernetwork 强度",
"Inpainting conditioning mask strength": "局部重绘时图像调节的蒙版屏蔽强度",
"Apply color correction to img2img results to match original colors.": "对图生图结果应用颜色校正以匹配原始颜色",
"Save a copy of image before applying color correction to img2img results": "在对图生图结果应用颜色校正之前保存图像副本",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "在进行图生图的时候,确切地执行滑块指定的迭代步数(正常情况下更弱的去噪需要更少的迭代步数)",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "在进行图生图的时候,确切地执行滑块指定的迭代步数(正常情况下更弱的重绘幅度需要更少的迭代步数)",
"Enable quantization in K samplers for sharper and cleaner results. This may change existing seeds. Requires restart to apply.": "在 K 采样器中启用量化以获得更清晰、更清晰的结果。这可能会改变现有的随机种子。需要重新启动才能应用",
"Emphasis: use (text) to make model pay more attention to text and [text] to make it pay less attention": "强调符:使用 (文字) 使模型更关注该文本,使用 [文字] 使其减少关注",
"Use old emphasis implementation. Can be useful to reproduce old seeds.": "使用旧的强调符实现。可用于复现旧随机种子",
"Make K-diffusion samplers produce same images in a batch as when making a single image": "使 K-diffusion 采样器批量生成与生成单个图像时产出相同的图像",
"Make K-diffusion samplers produce same images in a batch as when making a single image": "使 K-diffusion 采样器 批量生成与生成单个图像时产出相同的图像",
"Increase coherency by padding from the last comma within n tokens when using more than 75 tokens": "当使用超过 75 个 token 时,通过从 n 个 token 中的最后一个逗号填补来提高一致性",
"Filter NSFW content": "过滤成人内容",
"Stop At last layers of CLIP model": "在 CLIP 模型的最后哪一层停下",
"Filter NSFW content": "过滤成人内容(NSFW)",
"Stop At last layers of CLIP model": "在 CLIP 模型的最后哪一层停下 (Clip skip)",
"Interrogate Options": "反推提示词选项",
"Interrogate: keep models in VRAM": "反推: 将模型保存在显存(VRAM)中",
"Interrogate: use artists from artists.csv": "反推: 使用 artists.csv 中的艺术家",
@ -384,105 +418,207 @@
"User interface": "用户界面",
"Show progressbar": "显示进度条",
"Show image creation progress every N sampling steps. Set 0 to disable.": "每 N 个采样迭代步数显示图像生成进度。设置 0 禁用",
"Show previews of all images generated in a batch as a grid": "以网格的形式预览所有批量生成出来的图像",
"Show grid in results for web": "在网页的结果中显示概览图",
"Show previews of all images generated in a batch as a grid": "以网格的形式预览批量生成的所有图像",
"Show grid in results for web": "在网页的结果中显示宫格图",
"Do not show any images in results for web": "不在网页的结果中显示任何图像",
"Add model hash to generation information": "将模型的哈希值添加到生成信息",
"Add model name to generation information": "将模型名称添加到生成信息",
"When reading generation parameters from text into UI (from PNG info or pasted text), do not change the selected model/checkpoint.": "当从文本读取生成参数到 UI从 PNG 信息或粘贴文本)时,不要更改选定的模型(ckpt)",
"Font for image grids that have text": "有文字的概览图使用的字体",
"When reading generation parameters from text into UI (from PNG info or pasted text), do not change the selected model/checkpoint.": "从文本读取生成参数到用户界面(从 PNG 图片信息或粘贴文本)时,不要更改选定的模型(ckpt)",
"Send seed when sending prompt or image to other interface": "将提示词或者图片发送到 >> 其他界面时,把随机种子也传送过去",
"Font for image grids that have text": "有文字的宫格图使用的字体",
"Enable full page image viewer": "启用整页图像查看器",
"Show images zoomed in by default in full page image viewer": "在整页图像查看器中默认放大显示图像",
"Show images zoomed in by default in full page image viewer": "在整页图像查看器中默认放大显示图像",
"Show generation progress in window title.": "在窗口标题中显示生成进度",
"Quicksettings list": "快速设置列表",
"Localization (requires restart)": "本地化(需要重新启动",
"Localization (requires restart)": "本地化翻译需要保存设置并重启Gradio",
"Sampler parameters": "采样器参数",
"Hide samplers in user interface (requires restart)": "在用户界面中隐藏采样器(需要重新启动)",
"eta (noise multiplier) for DDIM": "DDIM 的 eta (噪声乘数) ",
"eta (noise multiplier) for ancestral samplers": "ancestral 采样器的 eta (噪声乘数)",
"img2img DDIM discretize": "图生图 DDIM 离散化",
"uniform": "均",
"uniform": "均",
"quad": "二阶",
"sigma churn": "sigma churn",
"sigma tmin": "最小(tmin) sigma",
"sigma noise": "sigma 噪声",
"Eta noise seed delta": "Eta 噪声种子偏移(noise seed delta)",
"Eta noise seed delta": "Eta 噪声种子偏移(ENSD - Eta noise seed delta)",
"Images Browser": "图库浏览器",
"Preload images at startup": "在启动时预载图像",
"Preload images at startup": "在启动时预载图像",
"Number of columns on the page": "每页列数",
"Number of rows on the page": "每页行数",
"Number of pictures displayed on each page": "每页显示的图像数量",
"Minimum number of pages per load": "每次加载的最小页数",
"Number of grids in each row": "每行显示多少格",
"Wildcards": "通配符",
"Use same seed for all images": "为所有图像使用同一个随机种子",
"Request browser notifications": "请求浏览器通知",
"Download localization template": "下载本地化模板",
"Reload custom script bodies (No ui updates, No restart)": "重新加载自定义脚本主体(无 ui 更新,无重启)",
"Reload custom script bodies (No ui updates, No restart)": "重新加载自定义脚本主体(无用户界面更新,无重启)",
"Restart Gradio and Refresh components (Custom Scripts, ui.py, js and css only)": "重启 Gradio 及刷新组件仅限自定义脚本、ui.py、js 和 css",
"Prompt (press Ctrl+Enter or Alt+Enter to generate)": "提示词(按 Ctrl+Enter 或 Alt+Enter 生成)",
"Negative prompt (press Ctrl+Enter or Alt+Enter to generate)": "反向提示词(按 Ctrl+Enter 或 Alt+Enter 生成)",
"Add a random artist to the prompt.": "随机添加一个艺术家到提示词中",
"Read generation parameters from prompt or last generation if prompt is empty into user interface.": "从提示词中读取生成参数,如果提示词为空,则读取上一次的生成参数到用户界面",
"Save style": "储存为模版风格",
"Apply selected styles to current prompt": "将所选样式应用于当前提示",
"Stop processing current image and continue processing.": "停止处理当前图像并继续处理下一个",
"Stop processing images and return any results accumulated so far.": "停止处理图像并返回迄今为止累积的任何结果",
"Style to apply; styles have components for both positive and negative prompts and apply to both": "要应用的模版风格; 模版风格包含正向和反向提示词,并应用于两者",
"Available": "可用",
"Install from URL": "从网址安装",
"Apply and restart UI": "应用并重启用户界面",
"Check for updates": "检查更新",
"Extension": "扩展",
"URL": "网址",
"Update": "更新",
"a1111-sd-webui-tagcomplete": "Tag自动补全",
"unknown": "未知",
"deforum-for-automatic1111-webui": "Deforum",
"sd-dynamic-prompting": "动态提示词",
"stable-diffusion-webui-aesthetic-gradients": "美术风格梯度",
"stable-diffusion-webui-aesthetic-image-scorer": "美术风格评分",
"stable-diffusion-webui-artists-to-study": "艺术家图库",
"stable-diffusion-webui-dataset-tag-editor": "数据集标签编辑器",
"stable-diffusion-webui-images-browser": "图库浏览器",
"stable-diffusion-webui-inspiration": "灵感",
"stable-diffusion-webui-wildcards": "通配符",
"Load from:": "加载自",
"Extension index URL": "扩展列表链接",
"URL for extension's git repository": "扩展的 git 仓库链接",
"Local directory name": "本地路径名",
"Prompt (press Ctrl+Enter or Alt+Enter to generate)": "提示词(按 Ctrl+Enter 或 Alt+Enter 生成)\nPrompt",
"Negative prompt (press Ctrl+Enter or Alt+Enter to generate)": "反向提示词(按 Ctrl+Enter 或 Alt+Enter 生成)\nNegative prompt",
"Stop processing current image and continue processing.": "停止处理当前图像,并继续处理下一个",
"Stop processing images and return any results accumulated so far.": "停止处理图像,并返回迄今为止累积的任何结果",
"Do not do anything special": "什么都不做",
"Which algorithm to use to produce the image": "使用哪种算法生成图像",
"Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help": "Euler Ancestral - 非常有创意,可以根据迭代步数获得完全不同的图像,将迭代步数设置为高于 30-40 不会有正面作用",
"Denoising Diffusion Implicit Models - best at inpainting": "Denoising Diffusion Implicit models - 最擅长内补绘制",
"Denoising Diffusion Implicit Models - best at inpainting": "Denoising Diffusion Implicit models - 最擅长局部重绘",
"Produce an image that can be tiled.": "生成可用于平铺(tiled)的图像",
"Use a two step process to partially create an image at smaller resolution, upscale, and then improve details in it without changing composition": "使用两步处理的时候以较小的分辨率生成初步图像、接着放大图像,然后在不更改构图的情况下改进其中的细节",
"Determines how little respect the algorithm should have for image's content. At 0, nothing will change, and at 1 you'll get an unrelated image. With values below 1.0, processing will take less steps than the Sampling Steps slider specifies.": "决定算法对图像内容的影响程度。设置 0 时,什么都不会改变,而在 1 时,你将获得不相关的图像。值低于 1.0 时,处理的迭代步数将少于“采样迭代步数”滑块指定的步数",
"Use a two step process to partially create an image at smaller resolution, upscale, and then improve details in it without changing composition": "使用两步处理的时候以较小的分辨率生成初步图像、接着放大图像,然后在不更改构图的情况下改进其中的细节",
"Determines how little respect the algorithm should have for image's content. At 0, nothing will change, and at 1 you'll get an unrelated image. With values below 1.0, processing will take less steps than the Sampling Steps slider specifies.": "决定算法对图像内容的影响程度。设置 0 时,什么都不会改变,而在 1 时,你将获得不相关的图像。\n值低于 1.0 时,处理的迭代步数将少于“采样迭代步数”滑块指定的步数",
"How many batches of images to create": "创建多少批次的图像",
"How many image to create in a single batch": "每批创建多少图像",
"Classifier Free Guidance Scale - how strongly the image should conform to prompt - lower values produce more creative results": "Classifier Free Guidance Scale - 图像应在多大程度上服从提示词 - 较低的值会产生更有创意的结果",
"A value that determines the output of random number generator - if you create an image with same parameters and seed as another image, you'll get the same result": "一个固定随机数生成器输出的值 - 以相同参数和随机种子生成的图像会得到相同的结果",
"Set seed to -1, which will cause a new random number to be used every time": "将随机种子设置为-1则每次都会使用一个新的随机数",
"Reuse seed from last generation, mostly useful if it was randomed": "重用上一次使用的随机种子,如果想要固定结果就会很有用",
"Seed of a different picture to be mixed into the generation.": "将要参与生成的另一张图的随机种子",
"How strong of a variation to produce. At 0, there will be no effect. At 1, you will get the complete picture with variation seed (except for ancestral samplers, where you will just get something).": "想要产生多强烈的变化。设为 0 时,将没有效果。设为 1 时你将获得完全产自差异随机种子的图像ancestral 采样器除外,你只是单纯地生成了一些东西)",
"Make an attempt to produce a picture similar to what would have been produced with same seed at specified resolution": "尝试生成与在指定分辨率下使用相同随机种子生成的图像相似的图片",
"This text is used to rotate the feature space of the imgs embs": "此文本用于旋转图集 embeddings 的特征空间",
"Separate values for X axis using commas.": "使用逗号分隔 X 轴的值",
"Separate values for Y axis using commas.": "使用逗号分隔 Y 轴的值",
"Write image to a directory (default - log/images) and generation parameters into csv file.": "将图像写入目录(默认 - log/images并将生成参数写入 csv 文件",
"Open images output directory": "打开图像输出目录",
"Write image to a directory (default - log/images) and generation parameters into csv file.": "将图像写入目录(默认 - log/images并将生成参数写入 csv 表格文件",
"How much to blur the mask before processing, in pixels.": "处理前要对蒙版进行多强的模糊,以像素为单位",
"What to put inside the masked area before processing it with Stable Diffusion.": "在使用 Stable Diffusion 处理蒙版区域之前要在蒙版区域内放置什么",
"fill it with colors of the image": "用图像的颜色填充它",
"keep whatever was there originally": "保留原来的东西",
"fill it with latent space noise": "用潜空间的噪声填充它",
"fill it with latent space zeroes": "用潜空间的零填充它",
"Upscale masked region to target resolution, do inpainting, downscale back and paste into original image": "将蒙版区域放大到目标分辨率,做内补绘制,缩小后粘贴到原始图像中",
"fill it with colors of the image": "用图像的颜色(高强度模糊)填充它",
"keep whatever was there originally": "保留原来的图像,不进行预处理",
"fill it with latent space noise": "于潜空间填充噪声",
"fill it with latent space zeroes": "于潜空间填零",
"Upscale masked region to target resolution, do inpainting, downscale back and paste into original image": "将蒙版区域(包括预留像素长度的缓冲区域)放大到目标分辨率,进行局部重绘。\n然后缩小并粘贴回原始图像中",
"Resize image to target resolution. Unless height and width match, you will get incorrect aspect ratio.": "将图像大小调整为目标分辨率。除非高度和宽度匹配,否则你将获得不正确的纵横比",
"Resize the image so that entirety of target resolution is filled with the image. Crop parts that stick out.": "调整图像大小,使整个目标分辨率都被图像填充。裁剪多出来的部分",
"Resize the image so that entirety of image is inside target resolution. Fill empty space with image's colors.": "调整图像大小,使整个图像在目标分辨率内。用图像的颜色填充空白区域",
"How many times to repeat processing an image and using it as input for the next iteration": "重复处理图像并用作下次迭代输入的次数",
"In loopback mode, on each loop the denoising strength is multiplied by this value. <1 means decreasing variety so your sequence will converge on a fixed picture. >1 means increasing variety so your sequence will become more and more chaotic.": "在回送模式下,在每个循环中,去噪强度都会乘以该值。<1 表示减少多样性,因此你的这一组图将集中在固定的图像上。>1 意味着增加多样性,因此你的这一组图将变得越来越混乱",
"In loopback mode, on each loop the denoising strength is multiplied by this value. <1 means decreasing variety so your sequence will converge on a fixed picture. >1 means increasing variety so your sequence will become more and more chaotic.": "在回送模式下,在每个循环中,重绘幅度都会乘以该值。<1 表示减少多样性,因此你的这一组图将集中在固定的图像上。>1 意味着增加多样性,因此你的这一组图将变得越来越混乱",
"For SD upscale, how much overlap in pixels should there be between tiles. Tiles overlap so that when they are merged back into one picture, there is no clearly visible seam.": "使用 SD 放大(SD upscale)时,图块(Tiles)之间应该有多少像素重叠。图块(Tiles)之间需要重叠才可以让它们在合并回一张图像时,没有清晰可见的接缝",
"A directory on the same machine where the server is running.": "与服务器主机上的目录",
"Leave blank to save images to the default path.": "留空以将图像保存到默认路径",
"Result = A * (1 - M) + B * M": "结果 = A * (1 - M) + B * M",
"Result = A + (B - C) * M": "结果 = A + (B - C) * M",
"1st and last digit must be 1. ex:'1, 2, 1'": "第一个和最后一个数字必须是 1。例:'1, 2, 1'",
"how fast should the training go. Low values will take longer to train, high values may fail to converge (not generate accurate results) and/or may break the embedding (This has happened if you see Loss: nan in the training info textbox. If this happens, you need to manually restore your embedding from an older not-broken backup).\n\nYou can set a single numeric value, or multiple learning rates using the syntax:\n\n rate_1:max_steps_1, rate_2:max_steps_2, ...\n\nEG: 0.005:100, 1e-3:1000, 1e-5\n\nWill train with rate of 0.005 for first 100 steps, then 1e-3 until 1000 steps, then 1e-5 for all remaining steps.": "训练应该多快。低值将需要更长的时间来训练,高值可能无法收敛(无法产生准确的结果)以及/也许可能会破坏 embedding如果你在训练信息文本框中看到 Loss: nan 就会发生这种情况。如果发生这种情况,你需要从较旧的未损坏的备份手动恢复 embedding\n\n你可以使用以下语法设置单个数值或多个学习率\n\n 率1:步限1, 率2:步限2, ...\n\n如: 0.005:100, 1e-3:1000, 1e-5\n\n即前 100 步将以 0.005 的速率训练,接着直到 1000 步为止以 1e-3 训练,然后剩余所有步以 1e-5 训练",
"Path to directory with input images": "带有输入图像的路径",
"Path to directory where to write outputs": "进行输出的路径",
"Input images directory": "输入图像目录",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; leave empty for default.": "使用以下标签定义如何选择图像的文件名: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; 默认请留空",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "使用以下标签定义如何选择图像的文件名: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; 默认请留空",
"If this option is enabled, watermark will not be added to created images. Warning: if you do not add watermark, you may be behaving in an unethical manner.": "如果启用此选项,水印将不会添加到生成出来的图像中。警告:如果你不添加水印,你的行为可能是不符合专业操守的",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; leave empty for default.": "使用以下标签定义如何选择图像和概览图的子目录: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; 默认请留空",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "使用以下标签定义如何选择图像和宫格图的子目录: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; 默认请留空",
"Restore low quality faces using GFPGAN neural network": "使用 GFPGAN 神经网络修复低质量面部",
"This regular expression will be used extract words from filename, and they will be joined using the option below into label text used for training. Leave empty to keep filename text as it is.": "此正则表达式将用于从文件名中提取单词,并将使用以下选项将它们接合到用于训练的标签文本中。留空以保持文件名文本不变",
"This string will be used to join split words into a single line if the option above is enabled.": "如果启用了上述选项,则此处的字符会用于将拆分的单词接合为同一行",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "设置名称列表,以逗号分隔,设置应转到顶部的快速访问栏,而不是通常的设置选项卡。有关设置名称,请参见 modules/shared.py。需要重新启动才能应用",
"Only applies to inpainting models. Determines how strongly to mask off the original image for inpainting and img2img. 1.0 means fully masked, which is the default behaviour. 0.0 means a fully unmasked conditioning. Lower values will help preserve the overall composition of the image, but will struggle with large changes.": "仅适用于局部重绘专用的模型(模型后缀为 inpainting.ckpt 的模型)。决定了蒙版在局部重绘以及图生图中屏蔽原图内容的强度。 1.0 表示完全屏蔽原图这是默认行为。0.0 表示完全不屏蔽让原图进行图像调节。较低的值将有助于保持原图的整体构图,但很难遇到较大的变化",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "设置项名称的列表,以逗号分隔,该设置会移动到顶部的快速访问栏,而不是默认的设置选项卡。有关设置名称,请参见 modules/shared.py。需要重新启动才能应用",
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "如果这个值不为零,它将被添加到随机种子中,并在使用带有 Eta 的采样器时用于初始化随机噪声。你可以使用它来产生更多的图像变化,或者你可以使用它来模仿其他软件生成的图像,如果你知道你在做什么",
"Leave empty for auto": "留空时自动生成",
"Autocomplete options": "自动补全选项",
"Enable Autocomplete": "开启Tag补全",
"Append commas": "附加逗号",
"latest": "最新",
"behind": "落后",
"Roll three": "抽三位出来",
"Generate forever": "无限生成",
"Cancel generate forever": "停止无限生成",
"how fast should the training go. Low values will take longer to train, high values may fail to converge (not generate accurate results) and/or may break the embedding (This has happened if you see Loss: nan in the training info textbox. If this happens, you need to manually restore your embedding from an older not-broken backup).\n\nYou can set a single numeric value, or multiple learning rates using the syntax:\n\n rate_1:max_steps_1, rate_2:max_steps_2, ...\n\nEG: 0.005:100, 1e-3:1000, 1e-5\n\nWill train with rate of 0.005 for first 100 steps, then 1e-3 until 1000 steps, then 1e-5 for all remaining steps.": "训练应该多快。低值将需要更长的时间来训练,高值可能无法收敛(无法产生准确的结果)以及/也许可能会破坏 embedding如果你在训练信息文本框中看到 Loss: nan 就会发生这种情况。如果发生这种情况,你需要从较旧的未损坏的备份手动恢复 embedding\n\n你可以使用以下语法设置单个数值或多个学习率\n\n 率1:步限1, 率2:步限2, ...\n\n如: 0.005:100, 1e-3:1000, 1e-5\n\n即前 100 步将以 0.005 的速率训练,接着直到 1000 步为止以 1e-3 训练,然后剩余所有步以 1e-5 训练",
"Move VAE and CLIP to RAM when training hypernetwork. Saves VRAM.": "训练时将 VAE 和 CLIP 从显存(VRAM)移放到内存(RAM),节省显存(VRAM)",
"How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results": "迭代改进生成的图像多少次;更高的值需要更长的时间;非常低的值会产生不好的结果",
"Draw a mask over an image, and the script will regenerate the masked area with content according to prompt": "在图像上画一个蒙版,脚本会根据提示重新生成蒙版区域的内容",
"Upscale image normally, split result into tiles, improve each tile using img2img, merge whole image back": "正常放大图像将结果分割成图块tiles用图生图改进每个图块tiles最后将整个图像合并回来",
"Create a grid where images will have different parameters. Use inputs below to specify which parameters will be shared by columns and rows": "创建一个网格,图像将有不同的参数。使用下面的输入来指定哪些参数将由列和行共享",
"Run Python code. Advanced user only. Must run program with --allow-code for this to work": "运行 Python 代码。仅限老手使用。必须以 --allow-code 来开启程序,才能使其运行",
"Separate a list of words with commas, and the first word will be used as a keyword: script will search for this word in the prompt, and replace it with others": "以逗号分割的单词列表,第一个单词将被用作关键词:脚本将在提示词中搜索这个单词,并用其他单词替换它",
"Separate a list of words with commas, and the script will make a variation of prompt with those words for their every possible order": "以逗号分割的单词列表,脚本会排列出这些单词的所有排列方式,并加入提示词各生成一次",
"Reconstruct prompt from existing image and put it into the prompt field.": "从现有的图像中重构出提示词,并将其放入提示词的输入文本框",
"Set the maximum number of words to be used in the [prompt_words] option; ATTENTION: If the words are too long, they may exceed the maximum length of the file path that the system can handle": "设置在[prompt_words]选项中要使用的最大字数;注意:如果字数太长,可能会超过系统可处理的文件路径的最大长度",
"Process an image, use it as an input, repeat.": "处理一张图像,将其作为输入,并重复",
"Insert selected styles into prompt fields": "在提示词中插入选定的模版风格",
"Save current prompts as a style. If you add the token {prompt} to the text, the style use that as placeholder for your prompt when you use the style in the future.": "将当前的提示词保存为模版风格。如果你在文本中添加{prompt}标记,那么将来你使用该模版风格时,你现有的提示词会替换模版风格中的{prompt}",
"Loads weights from checkpoint before making images. You can either use hash or a part of filename (as seen in settings) for checkpoint name. Recommended to use with Y axis for less switching.": "在生成图像之前从模型(ckpt)中加载权重。你可以使用哈希值或文件名的一部分(如设置中所示)作为模型(ckpt)名称。建议用在Y轴上以减少过程中模型的切换",
"Torch active: Peak amount of VRAM used by Torch during generation, excluding cached data.\nTorch reserved: Peak amount of VRAM allocated by Torch, including all active and cached data.\nSys VRAM: Peak amount of VRAM allocation across all applications / total GPU VRAM (peak utilization%).": "Torch active: 在生成过程中Torch使用的显存(VRAM)峰值,不包括缓存的数据。\nTorch reserved: Torch 分配的显存(VRAM)的峰值量,包括所有活动和缓存数据。\nSys VRAM: 所有应用程序分配的显存(VRAM)的峰值量 / GPU 的总显存(VRAM)(峰值利用率%",
"Uscale the image in latent space. Alternative is to produce the full image from latent representation, upscale that, and then move it back to latent space.": "放大潜空间中的图像。而另一种方法是,从潜变量表达中直接解码并生成完整的图像,接着放大它,然后再将其编码回潜空间",
"Start drawing": "开始绘制",
"Description": "描述",
"Action": "行动",
"Aesthetic Gradients": "美术风格梯度",
"aesthetic-gradients": "美术风格梯度",
"Wildcards": "通配符",
"Dynamic Prompts": "动态提示词",
"Image browser": "图库浏览器",
"images-browser": "图库浏览器",
"Inspiration": "灵感",
"Deforum": "Deforum",
"Artists to study": "艺术家图库",
"Aesthetic Image Scorer": "美术风格评分",
"Dataset Tag Editor": "数据集标签编辑器",
"----not work----": "----以下内容无法被翻译Bug----",
"Add a random artist to the prompt.": "随机添加一个艺术家到提示词中",
"Read generation parameters from prompt or last generation if prompt is empty into user interface.": "从提示词中读取生成参数,如果提示词为空,则读取上一次的生成参数到用户界面",
"Save style": "储存为模版风格",
"Apply selected styles to current prompt": "将所选模板风格,应用于当前提示词",
"Set seed to -1, which will cause a new random number to be used every time": "将随机种子设置为-1则每次都会使用一个新的随机数",
"Reuse seed from last generation, mostly useful if it was randomed": "重用上一次使用的随机种子,如果想要固定结果就会很有用",
"Open images output directory": "打开图像输出目录",
"Upscaler 1": "放大算法 1",
"Upscaler 2": "放大算法 2",
"Separate prompts into parts using vertical pipe character (|) and the script will create a picture for every combination of them (except for the first part, which will be present in all combinations)": "用竖线分隔符(|)将提示词分成若干部分,脚本将为它们的每一个组合创建一幅图片(除了被分割的第一部分,所有的组合都会包含这部分)",
"Select which Real-ESRGAN models to show in the web UI. (Requires restart)": "选择哪些 Real-ESRGAN 模型显示在网页用户界面。(需要重新启动)",
"Allowed categories for random artists selection when using the Roll button": "使用抽选艺术家按钮时将会随机的艺术家类别",
"Roll three": "抽三位出來",
"Generate forever": "不停地生成",
"Cancel generate forever": "取消不停地生成"
"Face restoration model": "面部修复模型",
"Install": "安装",
"Installing...": "安装中...",
"Installed": "已安装",
"Style to apply; styles have components for both positive and negative prompts and apply to both": "要使用的模版风格; 模版风格包含正向和反向提示词,并应用于两者\n\ud83c\udfa8 随机添加一个艺术家到提示词中\n \u2199\ufe0f 从提示词中读取生成参数,如果提示词为空,则读取上一次的生成参数到用户界面\n\ud83d\udcbe 将当前的提示词保存为模版风格(保存在styles.csv)\n\ud83d\udccb 将所选模板风格,应用于当前提示词\n如果你在文本中添加{prompt}标记,并保存为模版风格\n那么将来你使用该模版风格时你现有的提示词会替换模版风格中的{prompt}",
"A value that determines the output of random number generator - if you create an image with same parameters and seed as another image, you'll get the same result": "一个固定随机数生成器输出的值 - 以相同参数和随机种子生成的图像会得到相同的结果\n\ud83c\udfb2 将随机种子设置为-1则每次都会使用一个新的随机数\n\u267b\ufe0f 重用上一次使用的随机种子,如果想要固定输出结果就会很有用",
"----deprecated----": "----以下内容在webui新版本已移除----",
"▼": "▼",
"History": "历史记录",
"Show Textbox": "显示文本框",
"File with inputs": "含输入内容的文件",
"Prompts": "提示词",
"Disabled when launched with --hide-ui-dir-config.": "启动 --hide-ui-dir-config 时禁用",
"Open output directory": "打开输出目录",
"Create aesthetic images embedding": "生成美术风格图集 embedding",
"Split oversized images into two": "将过大的图像分为两份",
"Train an embedding; must specify a directory with a set of 1:1 ratio images": "训练 embedding 必须指定一组具有 1:1 比例图像的目录",
"Learning rate": "学习率",
"txt2img history": "文生图历史记录",
"img2img history": "图生图历史记录",
"extras history": "更多选项卡的历史记录",
"extras": "更多",
"custom fold": "自定义文件夹",
"Load": "载入",
"Prev batch": "上一批",
"Next batch": "下一批",
"number of images to delete consecutively next": "接下来要连续删除的图像数",
"Date to": "日期至",
"Refresh page": "刷新页面",
"Unload VAE and CLIP from VRAM when training": "训练时从显存(VRAM)中取消 VAE 和 CLIP 的加载",
"Number of pictures displayed on each page": "每页显示的图像数量",
"Number of grids in each row": "每行显示多少格",
"favorites": "收藏夹(已保存)",
"others": "其他",
"Collect": "收藏(保存)",
"--------": "--------"
}

View file

@ -7,19 +7,20 @@
"Loading...": "載入中…",
"view": "檢視",
"api": "api",
"•": "",
"•": "",
"built with gradio": "基於 Gradio 構建",
"Stable Diffusion checkpoint": "Stable Diffusion 模型權重存檔點",
"txt2img": "文生圖",
"img2img": "圖生圖",
"Extras": "後處理",
"PNG Info": "PNG 資訊",
"Checkpoint Merger": "模型權重存檔點合併工具",
"Extras": "更多",
"PNG Info": "圖片資訊",
"Checkpoint Merger": "模型權重存檔點合併",
"Train": "訓練",
"Create aesthetic embedding": "生成美術風格 embedding",
"Create aesthetic embedding": "生成美術風格",
"Image Browser": "圖庫瀏覽器",
"History": "歷史記錄",
"Settings": "設定",
"Extensions": "擴充",
"Prompt": "提示詞",
"Negative prompt": "反向提示詞",
"Run": "執行",
@ -28,7 +29,7 @@
"Generate": "生成",
"Style 1": "模版風格 1",
"Style 2": "模版風格 2",
"Label": "標",
"Label": "標",
"File": "檔案",
"Drop File Here": "拖曳檔案到此",
"-": "-",
@ -59,29 +60,29 @@
"Highres. fix": "高解析度修復",
"Firstpass width": "第一遍的寬度",
"Firstpass height": "第一遍的高度",
"Denoising strength": "去噪強度",
"Batch count": "批次",
"Batch size": "量",
"Denoising strength": "重繪幅度",
"Batch count": "生成批次",
"Batch size": "每批數量",
"CFG Scale": "提示詞相關性CFG",
"Seed": "隨機種子",
"Extra": "額外參數",
"Extra": "",
"Variation seed": "差異隨機種子",
"Variation strength": "差異強度",
"Resize seed from width": "自寬度縮放隨機種子",
"Resize seed from height": "自高度縮放隨機種子",
"Open for Clip Aesthetic!": "打開美術風格 Clip!",
"Open for Clip Aesthetic!": "打開以調整 Clip 的美術風格!",
"▼": "▼",
"Aesthetic weight": "美術風格權重",
"Aesthetic steps": "美術風格疊代步數",
"Aesthetic learning rate": "美術風格學習率",
"Slerp interpolation": "Slerp 插值",
"Slerp interpolation": "球面線性插值角度",
"Aesthetic imgs embedding": "美術風格圖集 embedding",
"None": "無",
"Aesthetic text for imgs": "該圖集的美術風格描述",
"Slerp angle": "Slerp 角度",
"Is negative text": "是反向提示詞",
"Script": "指令碼",
"Embedding to Shareable PNG": "將 Embedding 轉換為可分享的 PNG",
"Embedding to Shareable PNG": "將 Embedding 轉換為可分享的 PNG 圖片檔案",
"Prompt matrix": "提示詞矩陣",
"Prompts from file or textbox": "從文字方塊或檔案載入提示詞",
"X/Y plot": "X/Y 圖表",
@ -91,6 +92,10 @@
"Show Textbox": "顯示文字方塊",
"File with inputs": "含輸入內容的檔案",
"Prompts": "提示詞",
"Iterate seed every line": "每行輸入都換一個種子",
"Use same random seed for all lines": "每行輸入都使用同一個隨機種子",
"List of prompt inputs": "提示詞輸入列表",
"Upload prompt inputs": "上傳提示詞輸入檔案",
"X type": "X軸類型",
"Nothing": "無",
"Var. seed": "差異隨機種子",
@ -100,15 +105,16 @@
"Prompt order": "提示詞順序",
"Sampler": "採樣器",
"Checkpoint name": "模型權重存檔點的名稱",
"Hypernetwork": "Hypernetwork",
"Hypernet str.": "Hypernetwork 強度",
"Hypernetwork": "超網路(Hypernetwork",
"Hypernet str.": "超網路Hypernetwork強度",
"Sigma Churn": "Sigma Churn",
"Sigma min": "最小 Sigma",
"Sigma max": "最大 Sigma",
"Sigma noise": "Sigma noise",
"Eta": "Eta",
"Clip skip": "Clip 跳過",
"Denoising": "去噪",
"Denoising": "重繪幅度",
"Cond. Image Mask Weight": "圖像調節屏蔽度",
"X values": "X軸數值",
"Y type": "Y軸類型",
"Y values": "Y軸數值",
@ -118,44 +124,44 @@
"Drop Image Here": "拖曳圖像到此",
"Save": "儲存",
"Send to img2img": ">> 圖生圖",
"Send to inpaint": ">> 內補繪製",
"Send to extras": ">> 後處理",
"Send to inpaint": ">> 局部重繪",
"Send to extras": ">> 更多",
"Make Zip when Save?": "儲存時生成ZIP壓縮檔案",
"Textbox": "文字方塊",
"Interrogate\nCLIP": "CLIP\n反推提示詞",
"Interrogate\nDeepBooru": "DeepBooru\n反推提示詞",
"Inpaint": "內補繪製",
"Inpaint": "局部重繪",
"Batch img2img": "批量圖生圖",
"Image for img2img": "圖生圖的圖像",
"Image for inpainting with mask": "用於內補繪製蒙版內容的圖像",
"Image for inpainting with mask": "用於局部重繪並手動畫蒙版的圖像",
"Mask": "蒙版",
"Mask blur": "蒙版模糊",
"Mask mode": "蒙版模式",
"Draw mask": "繪製蒙版",
"Upload mask": "上傳蒙版",
"Masking mode": "蒙版模式",
"Inpaint masked": "內補繪製蒙版內容",
"Inpaint not masked": "內補繪製非蒙版內容",
"Inpaint masked": "重繪蒙版內容",
"Inpaint not masked": "重繪非蒙版內容",
"Masked content": "蒙版蒙住的內容",
"fill": "填充",
"original": "原圖",
"latent noise": "潛空間噪聲",
"latent nothing": "潛空間數值零",
"Inpaint at full resolution": "以完整解析度進行內補繪製",
"Inpaint at full resolution padding, pixels": "以完整解析度進行內補繪製 — 填補畫素",
"Process images in a directory on the same machine where the server is running.": "在伺服器主機上的目錄中處理圖像",
"Use an empty output directory to save pictures normally instead of writing to the output directory.": "指定一個空的資料夾為輸出目錄而非預設的 output 資料夾為輸出目錄",
"Inpaint at full resolution": "全解析度局部重繪",
"Inpaint at full resolution padding, pixels": "預留畫素",
"Process images in a directory on the same machine where the server is running.": "使用伺服器主機上的一個目錄,作為輸入目錄處理圖像",
"Use an empty output directory to save pictures normally instead of writing to the output directory.": "使用一個空的資料夾作為輸出目錄,而不是使用預設的 output 資料夾作為輸出目錄",
"Disabled when launched with --hide-ui-dir-config.": "啟動 --hide-ui-dir-config 時禁用",
"Input directory": "輸入目錄",
"Output directory": "輸出目錄",
"Resize mode": "縮放模式",
"Just resize": "只縮放",
"Crop and resize": "縮放並剪裁",
"Resize and fill": "縮放並填充",
"Just resize": "拉伸",
"Crop and resize": "",
"Resize and fill": "填充",
"img2img alternative test": "圖生圖的另一種測試",
"Loopback": "回送",
"Outpainting mk2": "繪製第二版",
"Poor man's outpainting": "效果稍差的繪製",
"Outpainting mk2": "外繪製第二版",
"Poor man's outpainting": "效果稍差的外繪製",
"SD upscale": "使用 SD 放大",
"should be 2 or lower.": "必須小於等於2",
"Override `Sampling method` to Euler?(this method is built for it)": "覆寫「採樣方法」為 Euler這個方法就是為這樣做設計的",
@ -164,15 +170,15 @@
"Original negative prompt": "初始反向提示詞",
"Override `Sampling Steps` to the same value as `Decode steps`?": "覆寫「採樣疊代步數」為「解碼疊代步數」?",
"Decode steps": "解碼疊代步數",
"Override `Denoising strength` to 1?": "覆寫「去噪強度」為1?",
"Override `Denoising strength` to 1?": "覆寫「重繪幅度」為1?",
"Decode CFG scale": "解碼提示詞相關性CFG",
"Randomness": "隨機度",
"Sigma adjustment for finding noise for image": "為尋找圖中噪點的 Sigma 調整",
"Loops": "疊代次數",
"Denoising strength change factor": "去噪強度的調整係數",
"Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8": "推薦設定採樣疊代步數80-100採樣器Euler a去噪強0.8",
"Denoising strength change factor": "重繪幅度的調整係數",
"Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8": "推薦設定採樣疊代步數80-100採樣器Euler a重繪幅0.8",
"Pixels to expand": "拓展的畫素數",
"Outpainting direction": "繪製的方向",
"Outpainting direction": "外繪製的方向",
"left": "左",
"right": "右",
"up": "上",
@ -183,6 +189,7 @@
"Tile overlap": "圖塊重疊的畫素",
"Upscaler": "放大演算法",
"Lanczos": "Lanczos",
"Nearest": "最鄰近(整數縮放)",
"LDSR": "LDSR",
"BSRGAN 4x": "BSRGAN 4x",
"ESRGAN_4x": "ESRGAN_4x",
@ -205,6 +212,7 @@
"CodeFormer visibility": "CodeFormer 可見度",
"CodeFormer weight (0 = maximum effect, 1 = minimum effect)": "CodeFormer 權重 0 = 最大效果, 1 = 最小效果)",
"Open output directory": "打開輸出目錄",
"Upscale Before Restoring Faces": "放大後再進行面部修復",
"Send to txt2img": ">> 文生圖",
"A merger of the two checkpoints will be generated in your": "合併後的模型權重存檔點會生成在你的",
"checkpoint": "模型權重存檔點",
@ -219,27 +227,57 @@
"Add difference": "加入差分",
"Save as float16": "以 float16 儲存",
"See": "檢視",
"wiki": "wiki",
"wiki": "wiki文件",
"for detailed explanation.": "以了解詳細說明",
"Create embedding": "生成 embedding",
"Create aesthetic images embedding": "生成美術風格圖集 embedding",
"Create hypernetwork": "生成 hypernetwork",
"Create hypernetwork": "生成超網路Hypernetwork",
"Preprocess images": "圖像預處理",
"Name": "名稱",
"Initialization text": "初始化文字",
"Number of vectors per token": "每個 token 的向量數",
"Overwrite Old Embedding": "覆寫舊的 Embedding",
"Modules": "模組",
"Enter hypernetwork layer structure": "輸入 hypernetwork 層結構",
"Select activation function of hypernetwork": "選擇 hypernetwork 的激活函數",
"Enter hypernetwork layer structure": "輸入超網路Hypernetwork層結構",
"Select activation function of hypernetwork": "選擇超網路Hypernetwork的激活函數",
"linear": "linear",
"relu": "relu",
"leakyrelu": "leakyrelu",
"elu": "elu",
"swish": "swish",
"tanh": "tanh",
"sigmoid": "sigmoid",
"celu": "celu",
"gelu": "gelu",
"glu": "glu",
"hardshrink": "hardshrink",
"hardsigmoid": "hardsigmoid",
"hardtanh": "hardtanh",
"logsigmoid": "logsigmoid",
"logsoftmax": "logsoftmax",
"mish": "mish",
"prelu": "prelu",
"rrelu": "rrelu",
"relu6": "relu6",
"selu": "selu",
"silu": "silu",
"softmax": "softmax",
"softmax2d": "softmax2d",
"softmin": "softmin",
"softplus": "softplus",
"softshrink": "softshrink",
"softsign": "softsign",
"tanhshrink": "tanhshrink",
"threshold": "閾值",
"Select Layer weights initialization. relu-like - Kaiming, sigmoid-like - Xavier is recommended": "挑選初始化層權重的方案. 類relu - Kaiming, 類sigmoid - Xavier 都是比較推薦的選項",
"Normal": "正態",
"KaimingUniform": "Kaiming 均勻",
"KaimingNormal": "Kaiming 正態",
"XavierUniform": "Xavier 均勻",
"XavierNormal": "Xavier 正態",
"Add layer normalization": "加入層標準化",
"Use dropout": "採用 dropout 防止過擬合",
"Overwrite Old Hypernetwork": "覆寫舊的 Hypernetwork",
"Overwrite Old Hypernetwork": "覆寫舊的超網路Hypernetwork",
"Source directory": "來源目錄",
"Destination directory": "目標目錄",
"Existing Caption txt Action": "對已有的TXT說明文字的行為",
@ -250,17 +288,22 @@
"Create flipped copies": "生成鏡像副本",
"Split oversized images into two": "將過大的圖像分為兩份",
"Split oversized images": "分割過大的圖像",
"Auto focal point crop": "自動焦點裁切",
"Use BLIP for caption": "使用 BLIP 生成說明文字(自然語言描述)",
"Use deepbooru for caption": "使用 deepbooru 生成說明文字(標",
"Use deepbooru for caption": "使用 deepbooru 生成說明文字(標",
"Split image threshold": "圖像分割閾值",
"Split image overlap ratio": "分割圖像重疊的比率",
"Focal point face weight": "焦點面部權重",
"Focal point entropy weight": "焦點熵權重",
"Focal point edges weight": "焦點線條權重",
"Create debug image": "生成除錯圖片",
"Preprocess": "預處理",
"Train an embedding; must specify a directory with a set of 1:1 ratio images": "訓練 embedding 必須指定一組具有 1:1 比例圖像的目錄",
"Train an embedding or Hypernetwork; you must specify a directory with a set of 1:1 ratio images": "訓練 embedding 或者 hypernetwork 必須指定一組具有 1:1 比例圖像的目錄",
"[wiki]": "[wiki]",
"Train an embedding or Hypernetwork; you must specify a directory with a set of 1:1 ratio images": "訓練 embedding 或者超網路Hypernetwork 必須指定一組具有 1:1 比例圖像的目錄",
"[wiki]": "[wiki文件]",
"Embedding": "Embedding",
"Embedding Learning rate": "Embedding 學習率",
"Hypernetwork Learning rate": "Hypernetwork 學習率",
"Hypernetwork Learning rate": "超網路Hypernetwork學習率",
"Learning rate": "學習率",
"Dataset directory": "資料集目錄",
"Log directory": "日誌目錄",
@ -268,69 +311,83 @@
"Max steps": "最大疊代步數",
"Save an image to log directory every N steps, 0 to disable": "每 N 步儲存一個圖像到日誌目錄0 表示禁用",
"Save a copy of embedding to log directory every N steps, 0 to disable": "每 N 步將 embedding 的副本儲存到日誌目錄0 表示禁用",
"Save images with embedding in PNG chunks": "儲存圖像並在 PNG 檔案中嵌入 embedding 檔案",
"Read parameters (prompt, etc...) from txt2img tab when making previews": "進行預覽時從文生圖頁籤中讀取參數(提示詞等)",
"Train Hypernetwork": "訓練 Hypernetwork",
"Save images with embedding in PNG chunks": "儲存圖像並在 PNG 圖片檔案中嵌入 embedding 檔案",
"Read parameters (prompt, etc...) from txt2img tab when making previews": "進行預覽時從文生圖頁籤中讀取參數(提示詞等)",
"Train Hypernetwork": "訓練超網路Hypernetwork",
"Train Embedding": "訓練 Embedding",
"Create an aesthetic embedding out of any number of images": "從任意數量的圖像中建立美術風格 embedding",
"Create images embedding": "生成圖集 embedding",
"txt2img history": "文生圖歷史記錄",
"img2img history": "圖生圖歷史記錄",
"extras history": "後處理歷史記錄",
"Renew Page": "刷新頁面",
"extras": "後處理",
"favorites": "收藏夾",
"Favorites": "收藏夾",
"Others": "其他",
"custom fold": "自訂資料夾",
"Load": "載入",
"Images directory": "圖像目錄",
"Prev batch": "上一批",
"Next batch": "下一批",
"Dropdown": "下拉式清單",
"First Page": "首頁",
"Prev Page": "上一頁",
"Page Index": "頁數",
"Next Page": "下一頁",
"End Page": "尾頁",
"number of images to delete consecutively next": "接下來要連續刪除的圖像數",
"delete next": "刪除下一張",
"Delete": "刪除",
"sort by": "排序方式",
"path name": "路徑名",
"date": "日期",
"keyword": "搜尋",
"Generate Info": "生成資訊",
"File Name": "檔案名",
"Collect": "收藏",
"Refresh page": "刷新頁面",
"Date to": "日期至",
"Move to favorites": "移動到收藏夾",
"Renew Page": "刷新頁面",
"Number": "數量",
"set_index": "設定索引",
"load_switch": "載入開關",
"turn_page_switch": "翻頁開關",
"Checkbox": "核取方塊",
"Apply settings": "儲存設定",
"Saving images/grids": "儲存圖像/概覽圖",
"Saving images/grids": "儲存圖像/宮格圖",
"Always save all generated images": "始終儲存所有生成的圖像",
"File format for images": "圖像的檔案格式",
"Images filename pattern": "圖像檔案名格式",
"Always save all generated image grids": "始終儲存所有生成的概覽圖",
"File format for grids": "概覽圖的檔案格式",
"Add extended info (seed, prompt) to filename when saving grid": "儲存概覽時將擴展資訊(隨機種子,提示詞)加入到檔案名",
"Do not save grids consisting of one picture": "只有一張圖片時不要儲存概覽圖",
"Prevent empty spots in grid (when set to autodetect)": "(在自動檢測時)防止概覽圖中出現空位",
"Grid row count; use -1 for autodetect and 0 for it to be same as batch size": "概覽行數; 使用 -1 進行自動檢測,使用 0 使其與批量大小相同",
"Save text information about generation parameters as chunks to png files": "將有關生成參數的文本資訊作為塊儲存到PNG檔案中",
"Create a text file next to every image with generation parameters.": "儲存圖像時在每個圖像旁邊建立一個文本檔案儲存生成參數",
"Add number to filename when saving": "儲存的時候在檔案名里加入數字",
"Always save all generated image grids": "始終儲存所有生成的宮格圖",
"File format for grids": "宮格圖的檔案格式",
"Add extended info (seed, prompt) to filename when saving grid": "儲存宮格圖時,將擴展資訊(隨機種子,提示詞)加入到檔案名",
"Do not save grids consisting of one picture": "只有一張圖片時不要儲存宮格圖",
"Prevent empty spots in grid (when set to autodetect)": "(啟用自動偵測時)防止宮格圖中出現空位",
"Grid row count; use -1 for autodetect and 0 for it to be same as batch size": "宮格圖行數; 使用 -1 進行自動檢測,使用 0 使其與每批數量相同",
"Save text information about generation parameters as chunks to png files": "將有關生成參數的文本資訊作為塊儲存到PNG圖片檔案中",
"Create a text file next to every image with generation parameters.": "儲存圖像時,在每個圖像旁邊建立一個文本檔案儲存生成參數",
"Save a copy of image before doing face restoration.": "在進行面部修復之前儲存圖像副本",
"Save a copy of image before applying highres fix.": "在做高解析度修復之前儲存初始圖像副本",
"Save a copy of image before applying color correction to img2img results": "在對圖生圖結果套用顏色校正之前儲存圖像副本",
"Quality for saved jpeg images": "儲存的JPEG圖像的品質",
"If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG": "如果 PNG 圖像大於 4MB 或寬高大於 4000則縮小並儲存副本為 JPG",
"Use original name for output filename during batch process in extras tab": "在後處理頁籤中的批量處理過程中使用原始名稱作為輸出檔案名",
"If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG": "如果 PNG 圖像大於 4MB 或寬高大於 4000則縮小並儲存副本為 JPG 圖片",
"Use original name for output filename during batch process in extras tab": "在更多頁籤中的批量處理過程中,使用原始名稱作為輸出檔案名",
"When using 'Save' button, only save a single selected image": "使用「儲存」按鈕時,只儲存一個選定的圖像",
"Do not add watermark to images": "不要給圖像加浮水印",
"Paths for saving": "儲存路徑",
"Output directory for images; if empty, defaults to three directories below": "圖像的輸出目錄; 如果為空,則預設為以下三個目錄",
"Output directory for txt2img images": "文生圖的輸出目錄",
"Output directory for img2img images": "圖生圖的輸出目錄",
"Output directory for images from extras tab": "後處理的輸出目錄",
"Output directory for grids; if empty, defaults to two directories below": "概覽圖的輸出目錄; 如果為空,則預設為以下兩個目錄",
"Output directory for txt2img grids": "文生圖概覽的輸出目錄",
"Output directory for img2img grids": "圖生圖概覽的輸出目錄",
"Output directory for images from extras tab": "更多頁籤的輸出目錄",
"Output directory for grids; if empty, defaults to two directories below": "宮格圖的輸出目錄; 如果為空,則預設為以下兩個目錄",
"Output directory for txt2img grids": "文生圖宮格的輸出目錄",
"Output directory for img2img grids": "圖生圖宮格的輸出目錄",
"Directory for saving images using the Save button": "使用「儲存」按鈕儲存圖像的目錄",
"Saving to a directory": "儲存到目錄",
"Save images to a subdirectory": "將圖像儲存到子目錄",
"Save grids to a subdirectory": "將概覽圖儲存到子目錄",
"Save grids to a subdirectory": "將宮格圖儲存到子目錄",
"When using \"Save\" button, save images to a subdirectory": "使用「儲存」按鈕時,將圖像儲存到子目錄",
"Directory name pattern": "目錄名稱格式",
"Max prompt words for [prompt_words] pattern": "[prompt_words] 格式的最大提示詞數量",
@ -341,10 +398,10 @@
"Tile overlap, in pixels for SwinIR. Low values = visible seam.": "SwinIR 的圖塊重疊畫素。低值 = 可見接縫",
"LDSR processing steps. Lower = faster": "LDSR 處理疊代步數。更低 = 更快",
"Upscaler for img2img": "圖生圖的放大演算法",
"Upscale latent space image when doing hires. fix": "做高解析度修復時也放大潛空間圖像",
"Upscale latent space image when doing hires. fix": "做高解析度修復時也放大潛空間圖像",
"Face restoration": "面部修復",
"CodeFormer weight parameter; 0 = maximum effect; 1 = minimum effect": "CodeFormer 權重參數; 0 = 最大效果; 1 = 最小效果",
"Move face restoration model from VRAM into RAM after processing": "面部修復處理完成後將面部修復模型從顯存VRAM移至內存RAM",
"Move face restoration model from VRAM into RAM after processing": "面部修復處理完成後將面部修復模型從顯存VRAM移至內存RAM",
"System": "系統",
"VRAM usage polls per second during generation. Set to 0 to disable.": "生成圖像時每秒輪詢顯存VRAM使用情況的次數。設定為 0 以禁用",
"Always print all generation info to standard output": "始終將所有生成資訊輸出到 standard output (一般為控制台)",
@ -352,50 +409,55 @@
"Training": "訓練",
"Unload VAE and CLIP from VRAM when training": "訓練時從顯存VRAM中取消 VAE 和 CLIP 的載入",
"Move VAE and CLIP to RAM when training hypernetwork. Saves VRAM.": "訓練時將 VAE 和 CLIP 從顯存VRAM移放到內存RAM節省顯存VRAM",
"Move VAE and CLIP to RAM when training if possible. Saves VRAM.": "訓練時將 VAE 和 CLIP 從顯存VRAM移放到內存RAM如果可行的話節省顯存VRAM",
"Filename word regex": "檔案名用詞的正則表達式",
"Filename join string": "檔案名連接用字串",
"Number of repeats for a single input image per epoch; used only for displaying epoch number": "每個 epoch 中單個輸入圖像的重複次數; 僅用於顯示 epoch 數",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "每 N 步儲存一個包含 loss 的CSV到日誌目錄0 表示禁用",
"Save an csv containing the loss to log directory every N steps, 0 to disable": "每 N 步儲存一個包含 loss 的CSV表格到日誌目錄0 表示禁用",
"Use cross attention optimizations while training": "訓練時開啟 cross attention 最佳化",
"Stable Diffusion": "Stable Diffusion",
"Checkpoints to cache in RAM": "快取在內存RAM中的模型權重存檔點",
"Hypernetwork strength": "Hypernetwork 強度",
"SD VAE": "模型的VAE",
"auto": "自動",
"Hypernetwork strength": "超網路Hypernetwork強度",
"Inpainting conditioning mask strength": "局部重繪時圖像調節的蒙版屏蔽強度",
"Apply color correction to img2img results to match original colors.": "對圖生圖結果套用顏色校正以匹配原始顏色",
"Save a copy of image before applying color correction to img2img results": "在對圖生圖結果套用顏色校正之前儲存圖像副本",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "在進行圖生圖的時候,確切地執行滑塊指定的疊代步數(正常情況下更弱的去噪需要更少的疊代步數)",
"With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising).": "在進行圖生圖的時候,確切地執行滑塊指定的疊代步數(正常情況下更弱的重繪幅度需要更少的疊代步數)",
"Enable quantization in K samplers for sharper and cleaner results. This may change existing seeds. Requires restart to apply.": "在 K 採樣器中啟用量化以獲得更清晰,更清晰的結果。這可能會改變現有的隨機種子。需要重新啟動才能套用",
"Emphasis: use (text) to make model pay more attention to text and [text] to make it pay less attention": "強調符:使用 (文字) 使模型更關注該文本,使用 [文字] 使其減少關注",
"Use old emphasis implementation. Can be useful to reproduce old seeds.": "使用舊的強調符實作。可用於復現舊隨機種子",
"Make K-diffusion samplers produce same images in a batch as when making a single image": "使 K-diffusion 採樣器批量生成與生成單個圖像時產出相同的圖像",
"Make K-diffusion samplers produce same images in a batch as when making a single image": "使 K-diffusion 採樣器批量生成與生成單個圖像時產出相同的圖像",
"Increase coherency by padding from the last comma within n tokens when using more than 75 tokens": "當使用超過 75 個 token 時,通過從 n 個 token 中的最後一個逗號填補來提高一致性",
"Filter NSFW content": "過濾成人內容",
"Stop At last layers of CLIP model": "在 CLIP 模型的最後哪一層停下",
"Interrogate Options": "反推提示詞選項",
"Interrogate: keep models in VRAM": "反推: 將模型儲存在顯存VRAM中",
"Interrogate: use artists from artists.csv": "反推: 使用 artists.csv 中的藝術家",
"Interrogate: include ranks of model tags matches in results (Has no effect on caption-based interrogators).": "反推: 在生成結果中包含與模型標相匹配的等級(對基於生成自然語言描述的反推沒有影響)",
"Interrogate: include ranks of model tags matches in results (Has no effect on caption-based interrogators).": "反推: 在生成結果中包含與模型標相匹配的等級(對基於生成自然語言描述的反推沒有影響)",
"Interrogate: num_beams for BLIP": "反推: BLIP 的 num_beams",
"Interrogate: minimum description length (excluding artists, etc..)": "反推: 最小描述長度(不包括藝術家, 等…)",
"Interrogate: minimum description length (excluding artists, etc..)": "反推: 最小描述長度(不包括藝術家等…)",
"Interrogate: maximum description length": "反推: 最大描述長度",
"CLIP: maximum number of lines in text file (0 = No limit)": "CLIP 文本檔案中的最大行數0 = 無限制)",
"Interrogate: deepbooru score threshold": "反推: deepbooru 分數閾值",
"Interrogate: deepbooru sort alphabetically": "反推: deepbooru 按字母順序排序",
"use spaces for tags in deepbooru": "在 deepbooru 中為標使用空格",
"use spaces for tags in deepbooru": "在 deepbooru 中為標使用空格",
"escape (\\) brackets in deepbooru (so they are used as literal brackets and not for emphasis)": "在 deepbooru 中使用轉義 (\\) 括號(因此它們用作文字括號而不是強調符號)",
"User interface": "使用者介面",
"Show progressbar": "顯示進度列",
"Show image creation progress every N sampling steps. Set 0 to disable.": "每 N 個採樣疊代步數顯示圖像生成進度。設定 0 禁用",
"Show previews of all images generated in a batch as a grid": "以網格的形式預覽所有批量生成出來的圖像",
"Show grid in results for web": "在網頁的結果中顯示概覽圖",
"Show previews of all images generated in a batch as a grid": "以網格的形式預覽批量生成的所有圖像",
"Show grid in results for web": "在網頁的結果中顯示宮格圖",
"Do not show any images in results for web": "不在網頁的結果中顯示任何圖像",
"Add model hash to generation information": "將模型的雜湊值加入到生成資訊",
"Add model name to generation information": "將模型名稱加入到生成資訊",
"When reading generation parameters from text into UI (from PNG info or pasted text), do not change the selected model/checkpoint.": "當從文本讀取生成參數到 UI從 PNG 資訊或粘貼文本)時,不要更改選定的模型權重存檔點",
"Font for image grids that have text": "有文字的概覽圖使用的字體",
"When reading generation parameters from text into UI (from PNG info or pasted text), do not change the selected model/checkpoint.": "從文本讀取生成參數到使用者介面(從 PNG 圖片資訊或粘貼文本)時,不要更改選定的模型權重存檔點",
"Send seed when sending prompt or image to other interface": "將提示詞或者圖片發送到 >> 其他界面時,把隨機種子也傳送過去",
"Font for image grids that have text": "有文字的宮格圖使用的字體",
"Enable full page image viewer": "啟用整頁圖像檢視器",
"Show images zoomed in by default in full page image viewer": "在整頁圖像檢視器中預設放大顯示圖像",
"Show images zoomed in by default in full page image viewer": "在整頁圖像檢視器中預設放大顯示圖像",
"Show generation progress in window title.": "在視窗標題中顯示生成進度",
"Quicksettings list": "快速設定列表",
"Localization (requires restart)": "本地化(需要重新啟動",
"Localization (requires restart)": "本地化翻譯需要儲存設定並重啟Gradio",
"Sampler parameters": "採樣器參數",
"Hide samplers in user interface (requires restart)": "在使用者介面中隱藏採樣器(需要重新啟動)",
"eta (noise multiplier) for DDIM": "DDIM 的 eta (噪聲乘數)",
@ -406,9 +468,9 @@
"sigma churn": "sigma churn",
"sigma tmin": "最小(tmin) sigma",
"sigma noise": "sigma 噪聲",
"Eta noise seed delta": "Eta 噪聲種子偏移(noise seed delta",
"Eta noise seed delta": "Eta 噪聲種子偏移(ENSD",
"Images Browser": "圖庫瀏覽器",
"Preload images at startup": "在啟動時預載圖像",
"Preload images at startup": "在啟動時預載圖像",
"Number of columns on the page": "每頁列數",
"Number of rows on the page": "每頁行數",
"Number of pictures displayed on each page": "每頁顯示的圖像數量",
@ -418,24 +480,46 @@
"Use same seed for all images": "為所有圖像使用同一個隨機種子",
"Request browser notifications": "請求瀏覽器通知",
"Download localization template": "下載本地化模板",
"Reload custom script bodies (No ui updates, No restart)": "重新載入自訂指令碼主體(無UI更新,無重啟)",
"Reload custom script bodies (No ui updates, No restart)": "重新載入自訂指令碼主體(無使用者介面更新,無重啟)",
"Restart Gradio and Refresh components (Custom Scripts, ui.py, js and css only)": "重啟 Gradio 及刷新組件僅限自訂指令碼ui.pyJS 和 CSS",
"Available": "可用",
"Install from URL": "從網址安裝",
"Apply and restart UI": "應用並重啟使用者介面",
"Check for updates": "檢查更新",
"Extension": "擴充",
"URL": "網址",
"Update": "更新",
"a1111-sd-webui-tagcomplete": "標記自動補全",
"unknown": "未知",
"deforum-for-automatic1111-webui": "Deforum",
"sd-dynamic-prompting": "動態提示詞",
"stable-diffusion-webui-aesthetic-gradients": "美術風格梯度",
"stable-diffusion-webui-aesthetic-image-scorer": "美術風格評等",
"stable-diffusion-webui-artists-to-study": "藝術家圖庫",
"stable-diffusion-webui-dataset-tag-editor": "資料集標記編輯器",
"stable-diffusion-webui-images-browser": "圖庫瀏覽器",
"stable-diffusion-webui-inspiration": "靈感",
"stable-diffusion-webui-wildcards": "萬用字元",
"Load from:": "載入自",
"Extension index URL": "擴充清單連結",
"URL for extension's git repository": "擴充的 git 倉庫連結",
"Local directory name": "本地路徑名",
"Prompt (press Ctrl+Enter or Alt+Enter to generate)": "提示詞(按 Ctrl+Enter 或 Alt+Enter 生成)",
"Negative prompt (press Ctrl+Enter or Alt+Enter to generate)": "反向提示詞(按 Ctrl+Enter 或 Alt+Enter 生成)",
"Add a random artist to the prompt.": "隨機加入一個藝術家到提示詞中",
"Read generation parameters from prompt or last generation if prompt is empty into user interface.": "從提示詞中讀取生成參數,如果提示詞為空,則讀取上一次的生成參數到使用者介面",
"Save style": "存儲為模板風格",
"Apply selected styles to current prompt": "將所選樣式套用於當前提示",
"Stop processing current image and continue processing.": "停止處理當前圖像並繼續處理下一個",
"Stop processing images and return any results accumulated so far.": "停止處理圖像並返回迄今為止累積的任何結果",
"Stop processing current image and continue processing.": "停止處理當前圖像並繼續處理下一個",
"Stop processing images and return any results accumulated so far.": "停止處理圖像並返回迄今為止累積的任何結果",
"Style to apply; styles have components for both positive and negative prompts and apply to both": "要套用的模版風格; 模版風格包含正向和反向提示詞,並套用於兩者",
"Do not do anything special": "什麼都不做",
"Which algorithm to use to produce the image": "使用哪種演算法生成圖像",
"Euler Ancestral - very creative, each can get a completely different picture depending on step count, setting steps to higher than 30-40 does not help": "Euler Ancestral - 非常有創意,可以根據疊代步數獲得完全不同的圖像,將疊代步數設定為高於 30-40 不會有正面作用",
"Denoising Diffusion Implicit Models - best at inpainting": "Denoising Diffusion Implicit models - 最擅長內補繪製",
"Denoising Diffusion Implicit Models - best at inpainting": "Denoising Diffusion Implicit models - 最擅長局部重繪",
"Produce an image that can be tiled.": "生成可用於平舖的圖像",
"Use a two step process to partially create an image at smaller resolution, upscale, and then improve details in it without changing composition": "使用兩步處理的時候以較小的解析度生成初步圖像,接著放大圖像,然後在不更改構圖的情況下改進其中的細節",
"Determines how little respect the algorithm should have for image's content. At 0, nothing will change, and at 1 you'll get an unrelated image. With values below 1.0, processing will take less steps than the Sampling Steps slider specifies.": "決定演算法對圖像內容的影響程度。設定 0 時,什麼都不會改變,而在 1 時,你將獲得不相關的圖像。值低於 1.0 時,處理的疊代步數將少於「採樣疊代步數」滑塊指定的步數",
"Use a two step process to partially create an image at smaller resolution, upscale, and then improve details in it without changing composition": "使用兩步處理的時候以較小的解析度生成初步圖像,接著放大圖像,然後在不更改構圖的情況下改進其中的細節",
"Determines how little respect the algorithm should have for image's content. At 0, nothing will change, and at 1 you'll get an unrelated image. With values below 1.0, processing will take less steps than the Sampling Steps slider specifies.": "決定演算法對圖像內容的影響程度。設定 0 時,什麼都不會改變,而在 1 時,你將獲得不相關的圖像。\n值低於 1.0 時,處理的疊代步數將少於「採樣疊代步數」滑塊指定的步數",
"How many batches of images to create": "建立多少批次的圖像",
"How many image to create in a single batch": "每批建立多少圖像",
"Classifier Free Guidance Scale - how strongly the image should conform to prompt - lower values produce more creative results": "Classifier Free Guidance Scale - 圖像應在多大程度上服從提示詞 - 較低的值會產生更有創意的結果",
@ -448,20 +532,20 @@
"This text is used to rotate the feature space of the imgs embs": "此文本用於旋轉圖集 embeddings 的特徵空間",
"Separate values for X axis using commas.": "使用逗號分隔 X 軸的值",
"Separate values for Y axis using commas.": "使用逗號分隔 Y 軸的值",
"Write image to a directory (default - log/images) and generation parameters into csv file.": "將圖像寫入目錄(預設 — log/images並將生成參數寫入CSV檔案",
"Write image to a directory (default - log/images) and generation parameters into csv file.": "將圖像寫入目錄(預設 — log/images並將生成參數寫入CSV表格檔案",
"Open images output directory": "打開圖像輸出目錄",
"How much to blur the mask before processing, in pixels.": "處理前要對蒙版進行多強的模糊,以畫素為單位",
"What to put inside the masked area before processing it with Stable Diffusion.": "在使用 Stable Diffusion 處理蒙版區域之前要在蒙版區域內放置什麼",
"fill it with colors of the image": "用圖像的顏色填充它",
"keep whatever was there originally": "保留原來的内容",
"fill it with latent space noise": "用潛空間的噪聲填充它",
"fill it with latent space zeroes": "用潛空間的零填充它",
"Upscale masked region to target resolution, do inpainting, downscale back and paste into original image": "將蒙版區域放大到目標解析度,做內補繪製,縮小後粘貼到原始圖像中",
"fill it with colors of the image": "用圖像的顏色(高強度模糊)填充它",
"keep whatever was there originally": "保留原來的圖像,不進行預處理",
"fill it with latent space noise": "於潛空間填充噪聲",
"fill it with latent space zeroes": "於潛空間填零",
"Upscale masked region to target resolution, do inpainting, downscale back and paste into original image": "將蒙版區域(包括預留畫素長度的緩衝區域)放大到目標解析度,進行局部重繪。\n然後縮小並粘貼回原始圖像中",
"Resize image to target resolution. Unless height and width match, you will get incorrect aspect ratio.": "將圖像大小調整為目標解析度。除非高度和寬度匹配,否則你將獲得不正確的縱橫比",
"Resize the image so that entirety of target resolution is filled with the image. Crop parts that stick out.": "調整圖像大小,使整個目標解析度都被圖像填充。裁剪多出來的部分",
"Resize the image so that entirety of image is inside target resolution. Fill empty space with image's colors.": "調整圖像大小,使整個圖像在目標解析度內。用圖像的顏色填充空白區域",
"How many times to repeat processing an image and using it as input for the next iteration": "重複處理圖像並用作下次疊代輸入的次數",
"In loopback mode, on each loop the denoising strength is multiplied by this value. <1 means decreasing variety so your sequence will converge on a fixed picture. >1 means increasing variety so your sequence will become more and more chaotic.": "在回送模式下,在每個循環中,去噪強度都會乘以該值。<1 表示減少多樣性,因此你的這一組圖將集中在固定的圖像上。>1 意味著增加多樣性,因此你的這一組圖將變得越來越混亂",
"In loopback mode, on each loop the denoising strength is multiplied by this value. <1 means decreasing variety so your sequence will converge on a fixed picture. >1 means increasing variety so your sequence will become more and more chaotic.": "在回送模式下,在每個循環中,重繪幅度都會乘以該值。<1 表示減少多樣性,因此你的這一組圖將集中在固定的圖像上。>1 意味著增加多樣性,因此你的這一組圖將變得越來越混亂",
"For SD upscale, how much overlap in pixels should there be between tiles. Tiles overlap so that when they are merged back into one picture, there is no clearly visible seam.": "使用 SD 放大時,圖塊之間應該有多少畫素重疊。圖塊之間需要重疊才可以讓它們在合併回一張圖像時,沒有清晰可見的接縫",
"A directory on the same machine where the server is running.": "與伺服器主機上的目錄",
"Leave blank to save images to the default path.": "留空以將圖像儲存到預設路徑",
@ -472,17 +556,55 @@
"Path to directory with input images": "帶有輸入圖像的路徑",
"Path to directory where to write outputs": "進行輸出的路徑",
"Input images directory": "輸入圖像目錄",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; leave empty for default.": "使用以下標定義如何選擇圖像的檔案名: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp] 預設請留空",
"Use following tags to define how filenames for images are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "使用以下標定義如何選擇圖像的檔案名: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp] 預設請留空",
"If this option is enabled, watermark will not be added to created images. Warning: if you do not add watermark, you may be behaving in an unethical manner.": "如果啟用此選項,浮水印將不會加入到生成出來的圖像中。警告:如果你不加入浮水印,你的行為可能是不符合道德操守的",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp]; leave empty for default.": "使用以下標籤定義如何選擇圖像和概覽圖的子目錄: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [job_timestamp] 預設請留空",
"Use following tags to define how subdirectories for images and grids are chosen: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp]; leave empty for default.": "使用以下標記定義如何選擇圖像和宮格圖的子目錄: [steps], [cfg], [prompt], [prompt_no_styles], [prompt_spaces], [width], [height], [styles], [sampler], [seed], [model_hash], [prompt_words], [date], [datetime], [datetime<Format>], [datetime<Format><Time Zone>], [job_timestamp] 預設請留空",
"Restore low quality faces using GFPGAN neural network": "使用 GFPGAN 神經網路修復低品質面部",
"This regular expression will be used extract words from filename, and they will be joined using the option below into label text used for training. Leave empty to keep filename text as it is.": "此正則表達式將用於從檔案名中提取單詞,並將使用以下選項將它們接合到用於訓練的標文本中。留空以保持檔案名文本不變",
"This regular expression will be used extract words from filename, and they will be joined using the option below into label text used for training. Leave empty to keep filename text as it is.": "此正則表達式將用於從檔案名中提取單詞,並將使用以下選項將它們接合到用於訓練的標文本中。留空以保持檔案名文本不變",
"This string will be used to join split words into a single line if the option above is enabled.": "如果啟用了上述選項,則此處的字元會用於將拆分的單詞接合為同一行",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "設定名稱列表,以逗號分隔,設定應轉到頂部的快速存取列,而不是通常的設定頁籤。有關設定名稱,請參見 modules/shared.py。需要重新啟動才能套用",
"Only applies to inpainting models. Determines how strongly to mask off the original image for inpainting and img2img. 1.0 means fully masked, which is the default behaviour. 0.0 means a fully unmasked conditioning. Lower values will help preserve the overall composition of the image, but will struggle with large changes.": "僅適用於局部重繪專用的模型(模型後綴為 inpainting.ckpt 的模型)。決定了蒙版在局部重繪以及圖生圖中屏蔽原圖內容的強度。 1.0 表示完全屏蔽原圖,這是預設行為。 0.0 表示完全不屏蔽讓原圖進行圖像調節。較低的值將有助於保持原圖的整體構圖,但很難遇到較大的變化",
"List of setting names, separated by commas, for settings that should go to the quick access bar at the top, rather than the usual setting tab. See modules/shared.py for setting names. Requires restarting to apply.": "設定項名稱列表,以逗號分隔,該設定會移動到頂部的快速存取列,而不是預設的設定頁籤。有關設定名稱,請參見 modules/shared.py。需要重新啟動才能套用",
"If this values is non-zero, it will be added to seed and used to initialize RNG for noises when using samplers with Eta. You can use this to produce even more variation of images, or you can use this to match images of other software if you know what you are doing.": "如果這個值不為零,它將被加入到隨機種子中,並在使用帶有 Eta 的採樣器時用於初始化隨機噪聲。你可以使用它來產生更多的圖像變化,或者你可以使用它來模仿其他軟體生成的圖像,如果你知道你在做什麼",
"Leave empty for auto": "留空時自動生成",
"Autocomplete options": "自動補全選項",
"Enable Autocomplete": "開啟Tag補全",
"Select which Real-ESRGAN models to show in the web UI. (Requires restart)": "選擇哪些 Real-ESRGAN 模型顯示在網頁使用者介面。(需要重新啟動)",
"Allowed categories for random artists selection when using the Roll button": "使用抽選藝術家按鈕時將會隨機的藝術家類別",
"Append commas": "附加逗號",
"latest": "最新",
"behind": "落後",
"Roll three": "抽三位出來",
"Generate forever": "不停地生成",
"Cancel generate forever": "取消不停地生成"
"Generate forever": "無限生成",
"Cancel generate forever": "停止無限生成",
"How many times to improve the generated image iteratively; higher values take longer; very low values can produce bad results": "疊代改進生成的圖像多少次;更高的值需要更長的時間;非常低的值會產生不好的結果",
"Draw a mask over an image, and the script will regenerate the masked area with content according to prompt": "在圖像上畫一個蒙版,指令碼會根據提示重新生成蒙版區域的內容",
"Upscale image normally, split result into tiles, improve each tile using img2img, merge whole image back": "正常放大圖像,將結果分割成圖塊,用圖生圖改進每個圖塊,最後將整個圖像合併回來",
"Create a grid where images will have different parameters. Use inputs below to specify which parameters will be shared by columns and rows": "創建一個網格,圖像將有不同的參數。使用下面的輸入來指定哪些參數將由列和行共享",
"Run Python code. Advanced user only. Must run program with --allow-code for this to work": "執行 Python 程式碼。僅限老手使用。必須以 --allow-code 來開啟程式,才能使其執行",
"Separate a list of words with commas, and the first word will be used as a keyword: script will search for this word in the prompt, and replace it with others": "以逗號分割的單詞列表,第一個單詞將被用作關鍵詞:指令碼將在提示詞中搜尋這個單詞,並用其他單詞替換它",
"Separate a list of words with commas, and the script will make a variation of prompt with those words for their every possible order": "以逗號分割的單詞列表,指令碼會排列出這些單詞的所有排列方式,並加入提示詞各生成一次",
"Reconstruct prompt from existing image and put it into the prompt field.": "從現有的圖像中重構出提示詞,並將其放入提示詞的輸入文字方塊",
"Set the maximum number of words to be used in the [prompt_words] option; ATTENTION: If the words are too long, they may exceed the maximum length of the file path that the system can handle": "設定在[prompt_words]選項中要使用的最大字數;注意:如果字數太長,可能會超過系統可處理的檔案路徑的最大長度",
"Process an image, use it as an input, repeat.": "處理一張圖像,將其作為輸入,並重複",
"Insert selected styles into prompt fields": "在提示詞中插入選定的模版風格",
"Save current prompts as a style. If you add the token {prompt} to the text, the style use that as placeholder for your prompt when you use the style in the future.": "將當前的提示詞儲存為模版風格。如果你在文本中加入{prompt}標記,那麼將來你使用該模版風格時,你現有的提示詞會替換模版風格中的{prompt}",
"Loads weights from checkpoint before making images. You can either use hash or a part of filename (as seen in settings) for checkpoint name. Recommended to use with Y axis for less switching.": "在生成圖像之前從模型權重存檔點中載入權重。你可以使用哈希值或檔案名的一部分如設定中所示作為模型權重存檔點名稱。建議用在Y軸上以減少過程中模型的切換",
"Torch active: Peak amount of VRAM used by Torch during generation, excluding cached data.\nTorch reserved: Peak amount of VRAM allocated by Torch, including all active and cached data.\nSys VRAM: Peak amount of VRAM allocation across all applications / total GPU VRAM (peak utilization%).": "Torch active: 在生成過程中Torch使用的顯存(VRAM)峰值,不包括快取的數據。\nTorch reserved: Torch 分配的顯存(VRAM)的峰值量,包括所有活動和快取數據。\nSys VRAM: 所有應用程式分配的顯存(VRAM)的峰值量 / GPU 的總顯存(VRAM)(峰值利用率%",
"Uscale the image in latent space. Alternative is to produce the full image from latent representation, upscale that, and then move it back to latent space.": "放大潛空間中的圖像。而另一種方法是,從潛變量表達中直接解碼並生成完整的圖像,接著放大它,然後再將其編碼回潛空間",
"Start drawing": "開始繪製",
"Description": "描述",
"Action": "行動",
"Aesthetic Gradients": "美術風格梯度",
"aesthetic-gradients": "美術風格梯度",
"Dynamic Prompts": "動態提示詞",
"images-browser": "圖庫瀏覽器",
"Inspiration": "靈感",
"Deforum": "Deforum",
"Artists to study": "藝術家圖庫",
"Aesthetic Image Scorer": "美術風格評分",
"Dataset Tag Editor": "數據集標記編輯器",
"Face restoration model": "面部修復模型",
"Install": "安裝",
"Installing...": "安裝中…",
"Installed": "已安裝"
}

View file

View file

@ -2,15 +2,18 @@ import base64
import io
import time
import uvicorn
from gradio.processing_utils import decode_base64_to_file, decode_base64_to_image
from fastapi import APIRouter, Depends, HTTPException
from threading import Lock
from gradio.processing_utils import encode_pil_to_base64, decode_base64_to_file, decode_base64_to_image
from fastapi import APIRouter, Depends, FastAPI, HTTPException
import modules.shared as shared
from modules import devices
from modules.api.models import *
from modules.processing import StableDiffusionProcessingTxt2Img, StableDiffusionProcessingImg2Img, process_images
from modules.sd_samplers import all_samplers
from modules.extras import run_extras, run_pnginfo
from PIL import PngImagePlugin
from modules.sd_models import checkpoints_list
from modules.realesrgan_model import get_realesrgan_models
from typing import List
def upscaler_to_index(name: str):
try:
@ -32,13 +35,25 @@ def setUpscalers(req: dict):
def encode_pil_to_base64(image):
buffer = io.BytesIO()
image.save(buffer, format="png")
return base64.b64encode(buffer.getvalue())
with io.BytesIO() as output_bytes:
# Copy any text-only metadata
use_metadata = False
metadata = PngImagePlugin.PngInfo()
for key, value in image.info.items():
if isinstance(key, str) and isinstance(value, str):
metadata.add_text(key, value)
use_metadata = True
image.save(
output_bytes, "PNG", pnginfo=(metadata if use_metadata else None)
)
bytes_data = output_bytes.getvalue()
return base64.b64encode(bytes_data)
class Api:
def __init__(self, app, queue_lock):
def __init__(self, app: FastAPI, queue_lock: Lock):
self.router = APIRouter()
self.app = app
self.queue_lock = queue_lock
@ -48,7 +63,20 @@ class Api:
self.app.add_api_route("/sdapi/v1/extra-batch-images", self.extras_batch_images_api, methods=["POST"], response_model=ExtrasBatchImagesResponse)
self.app.add_api_route("/sdapi/v1/png-info", self.pnginfoapi, methods=["POST"], response_model=PNGInfoResponse)
self.app.add_api_route("/sdapi/v1/progress", self.progressapi, methods=["GET"], response_model=ProgressResponse)
self.app.add_api_route("/sdapi/v1/interrogate", self.interrogateapi, methods=["POST"])
self.app.add_api_route("/sdapi/v1/interrupt", self.interruptapi, methods=["POST"])
self.app.add_api_route("/sdapi/v1/options", self.get_config, methods=["GET"], response_model=OptionsModel)
self.app.add_api_route("/sdapi/v1/options", self.set_config, methods=["POST"])
self.app.add_api_route("/sdapi/v1/cmd-flags", self.get_cmd_flags, methods=["GET"], response_model=FlagsModel)
self.app.add_api_route("/sdapi/v1/samplers", self.get_samplers, methods=["GET"], response_model=List[SamplerItem])
self.app.add_api_route("/sdapi/v1/upscalers", self.get_upscalers, methods=["GET"], response_model=List[UpscalerItem])
self.app.add_api_route("/sdapi/v1/sd-models", self.get_sd_models, methods=["GET"], response_model=List[SDModelItem])
self.app.add_api_route("/sdapi/v1/hypernetworks", self.get_hypernetworks, methods=["GET"], response_model=List[HypernetworkItem])
self.app.add_api_route("/sdapi/v1/face-restorers", self.get_face_restorers, methods=["GET"], response_model=List[FaceRestorerItem])
self.app.add_api_route("/sdapi/v1/realesrgan-models", self.get_realesrgan_models, methods=["GET"], response_model=List[RealesrganItem])
self.app.add_api_route("/sdapi/v1/prompt-styles", self.get_promp_styles, methods=["GET"], response_model=List[PromptStyleItem])
self.app.add_api_route("/sdapi/v1/artist-categories", self.get_artists_categories, methods=["GET"], response_model=List[str])
self.app.add_api_route("/sdapi/v1/artists", self.get_artists, methods=["GET"], response_model=List[ArtistItem])
def text2imgapi(self, txt2imgreq: StableDiffusionTxt2ImgProcessingAPI):
sampler_index = sampler_to_index(txt2imgreq.sampler_index)
@ -179,17 +207,96 @@ class Api:
progress = min(progress, 1)
shared.state.set_current_image()
current_image = None
if shared.state.current_image and not req.skip_current_image:
current_image = encode_pil_to_base64(shared.state.current_image)
return ProgressResponse(progress=progress, eta_relative=eta_relative, state=shared.state.dict(), current_image=current_image)
def interrogateapi(self, interrogatereq: InterrogateRequest):
image_b64 = interrogatereq.image
if image_b64 is None:
raise HTTPException(status_code=404, detail="Image not found")
img = self.__base64_to_image(image_b64)
# Override object param
with self.queue_lock:
processed = shared.interrogator.interrogate(img)
return InterrogateResponse(caption=processed)
def interruptapi(self):
shared.state.interrupt()
return {}
def get_config(self):
options = {}
for key in shared.opts.data.keys():
metadata = shared.opts.data_labels.get(key)
if(metadata is not None):
options.update({key: shared.opts.data.get(key, shared.opts.data_labels.get(key).default)})
else:
options.update({key: shared.opts.data.get(key, None)})
return options
def set_config(self, req: OptionsModel):
# currently req has all options fields even if you send a dict like { "send_seed": false }, which means it will
# overwrite all options with default values.
raise RuntimeError('Setting options via API is not supported')
reqDict = vars(req)
for o in reqDict:
setattr(shared.opts, o, reqDict[o])
shared.opts.save(shared.config_filename)
return
def get_cmd_flags(self):
return vars(shared.cmd_opts)
def get_samplers(self):
return [{"name":sampler[0], "aliases":sampler[2], "options":sampler[3]} for sampler in all_samplers]
def get_upscalers(self):
upscalers = []
for upscaler in shared.sd_upscalers:
u = upscaler.scaler
upscalers.append({"name":u.name, "model_name":u.model_name, "model_path":u.model_path, "model_url":u.model_url})
return upscalers
def get_sd_models(self):
return [{"title":x.title, "model_name":x.model_name, "hash":x.hash, "filename": x.filename, "config": x.config} for x in checkpoints_list.values()]
def get_hypernetworks(self):
return [{"name": name, "path": shared.hypernetworks[name]} for name in shared.hypernetworks]
def get_face_restorers(self):
return [{"name":x.name(), "cmd_dir": getattr(x, "cmd_dir", None)} for x in shared.face_restorers]
def get_realesrgan_models(self):
return [{"name":x.name,"path":x.data_path, "scale":x.scale} for x in get_realesrgan_models(None)]
def get_promp_styles(self):
styleList = []
for k in shared.prompt_styles.styles:
style = shared.prompt_styles.styles[k]
styleList.append({"name":style[0], "prompt": style[1], "negative_prompr": style[2]})
return styleList
def get_artists_categories(self):
return shared.artist_db.cats
def get_artists(self):
return [{"name":x[0], "score":x[1], "category":x[2]} for x in shared.artist_db.artists]
def launch(self, server_name, port):
self.app.include_router(self.router)
uvicorn.run(self.app, host=server_name, port=port)

View file

@ -1,11 +1,11 @@
import inspect
from click import prompt
from pydantic import BaseModel, Field, create_model
from typing import Any, Optional
from typing_extensions import Literal
from inflection import underscore
from modules.processing import StableDiffusionProcessingTxt2Img, StableDiffusionProcessingImg2Img
from modules.shared import sd_upscalers
from modules.shared import sd_upscalers, opts, parser
from typing import Dict, List
API_NOT_ALLOWED = [
"self",
@ -65,6 +65,7 @@ class PydanticModelGenerator:
self._model_name = model_name
self._class_data = merge_class_params(class_instance)
self._model_def = [
ModelDef(
field=underscore(k),
@ -109,12 +110,12 @@ StableDiffusionImg2ImgProcessingAPI = PydanticModelGenerator(
).generate_model()
class TextToImageResponse(BaseModel):
images: list[str] = Field(default=None, title="Image", description="The generated image in base64 format.")
images: List[str] = Field(default=None, title="Image", description="The generated image in base64 format.")
parameters: dict
info: str
class ImageToImageResponse(BaseModel):
images: list[str] = Field(default=None, title="Image", description="The generated image in base64 format.")
images: List[str] = Field(default=None, title="Image", description="The generated image in base64 format.")
parameters: dict
info: str
@ -131,6 +132,7 @@ class ExtrasBaseRequest(BaseModel):
upscaler_1: str = Field(default="None", title="Main upscaler", description=f"The name of the main upscaler to use, it has to be one of this list: {' , '.join([x.name for x in sd_upscalers])}")
upscaler_2: str = Field(default="None", title="Secondary upscaler", description=f"The name of the secondary upscaler to use, it has to be one of this list: {' , '.join([x.name for x in sd_upscalers])}")
extras_upscaler_2_visibility: float = Field(default=0, title="Secondary upscaler visibility", ge=0, le=1, allow_inf_nan=False, description="Sets the visibility of secondary upscaler, values should be between 0 and 1.")
upscale_first: bool = Field(default=False, title="Upscale first", description="Should the upscaler run before restoring faces?")
class ExtraBaseResponse(BaseModel):
html_info: str = Field(title="HTML info", description="A series of HTML tags containing the process info.")
@ -146,10 +148,10 @@ class FileData(BaseModel):
name: str = Field(title="File name")
class ExtrasBatchImagesRequest(ExtrasBaseRequest):
imageList: list[FileData] = Field(title="Images", description="List of images to work on. Must be Base64 strings")
imageList: List[FileData] = Field(title="Images", description="List of images to work on. Must be Base64 strings")
class ExtrasBatchImagesResponse(ExtraBaseResponse):
images: list[str] = Field(title="Images", description="The generated images in base64 format.")
images: List[str] = Field(title="Images", description="The generated images in base64 format.")
class PNGInfoRequest(BaseModel):
image: str = Field(title="Image", description="The base64 encoded PNG image")
@ -165,3 +167,75 @@ class ProgressResponse(BaseModel):
eta_relative: float = Field(title="ETA in secs")
state: dict = Field(title="State", description="The current state snapshot")
current_image: str = Field(default=None, title="Current image", description="The current image in base64 format. opts.show_progress_every_n_steps is required for this to work.")
class InterrogateRequest(BaseModel):
image: str = Field(default="", title="Image", description="Image to work on, must be a Base64 string containing the image's data.")
class InterrogateResponse(BaseModel):
caption: str = Field(default=None, title="Caption", description="The generated caption for the image.")
fields = {}
for key, value in opts.data.items():
metadata = opts.data_labels.get(key)
optType = opts.typemap.get(type(value), type(value))
if (metadata is not None):
fields.update({key: (Optional[optType], Field(
default=metadata.default ,description=metadata.label))})
else:
fields.update({key: (Optional[optType], Field())})
OptionsModel = create_model("Options", **fields)
flags = {}
_options = vars(parser)['_option_string_actions']
for key in _options:
if(_options[key].dest != 'help'):
flag = _options[key]
_type = str
if _options[key].default is not None: _type = type(_options[key].default)
flags.update({flag.dest: (_type,Field(default=flag.default, description=flag.help))})
FlagsModel = create_model("Flags", **flags)
class SamplerItem(BaseModel):
name: str = Field(title="Name")
aliases: List[str] = Field(title="Aliases")
options: Dict[str, str] = Field(title="Options")
class UpscalerItem(BaseModel):
name: str = Field(title="Name")
model_name: Optional[str] = Field(title="Model Name")
model_path: Optional[str] = Field(title="Path")
model_url: Optional[str] = Field(title="URL")
class SDModelItem(BaseModel):
title: str = Field(title="Title")
model_name: str = Field(title="Model Name")
hash: str = Field(title="Hash")
filename: str = Field(title="Filename")
config: str = Field(title="Config file")
class HypernetworkItem(BaseModel):
name: str = Field(title="Name")
path: Optional[str] = Field(title="Path")
class FaceRestorerItem(BaseModel):
name: str = Field(title="Name")
cmd_dir: Optional[str] = Field(title="Path")
class RealesrganItem(BaseModel):
name: str = Field(title="Name")
path: Optional[str] = Field(title="Path")
scale: Optional[int] = Field(title="Scale")
class PromptStyleItem(BaseModel):
name: str = Field(title="Name")
prompt: Optional[str] = Field(title="Prompt")
negative_prompt: Optional[str] = Field(title="Negative Prompt")
class ArtistItem(BaseModel):
name: str = Field(title="Name")
score: float = Field(title="Score")
category: str = Field(title="Category")

View file

@ -50,6 +50,7 @@ def mod2normal(state_dict):
def resrgan2normal(state_dict, nb=23):
# this code is copied from https://github.com/victorca25/iNNfer
if "conv_first.weight" in state_dict and "body.0.rdb1.conv1.weight" in state_dict:
re8x = 0
crt_net = {}
items = []
for k, v in state_dict.items():
@ -75,10 +76,18 @@ def resrgan2normal(state_dict, nb=23):
crt_net['model.3.bias'] = state_dict['conv_up1.bias']
crt_net['model.6.weight'] = state_dict['conv_up2.weight']
crt_net['model.6.bias'] = state_dict['conv_up2.bias']
crt_net['model.8.weight'] = state_dict['conv_hr.weight']
crt_net['model.8.bias'] = state_dict['conv_hr.bias']
crt_net['model.10.weight'] = state_dict['conv_last.weight']
crt_net['model.10.bias'] = state_dict['conv_last.bias']
if 'conv_up3.weight' in state_dict:
# modification supporting: https://github.com/ai-forever/Real-ESRGAN/blob/main/RealESRGAN/rrdbnet_arch.py
re8x = 3
crt_net['model.9.weight'] = state_dict['conv_up3.weight']
crt_net['model.9.bias'] = state_dict['conv_up3.bias']
crt_net[f'model.{8+re8x}.weight'] = state_dict['conv_hr.weight']
crt_net[f'model.{8+re8x}.bias'] = state_dict['conv_hr.bias']
crt_net[f'model.{10+re8x}.weight'] = state_dict['conv_last.weight']
crt_net[f'model.{10+re8x}.bias'] = state_dict['conv_last.bias']
state_dict = crt_net
return state_dict

View file

@ -34,8 +34,11 @@ class Extension:
if repo is None or repo.bare:
self.remote = None
else:
self.remote = next(repo.remote().urls, None)
self.status = 'unknown'
try:
self.remote = next(repo.remote().urls, None)
self.status = 'unknown'
except Exception:
self.remote = None
def list_files(self, subdir, extension):
from modules import scripts

View file

@ -136,12 +136,13 @@ def run_extras(extras_mode, resize_mode, image, image_folder, input_dir, output_
def run_upscalers_blend(params: List[UpscaleParams], image: Image.Image, info: str) -> Tuple[Image.Image, str]:
blended_result: Image.Image = None
image_hash: str = hash(np.array(image.getdata()).tobytes())
for upscaler in params:
upscale_args = (upscaler.upscaler_idx, upscaling_resize, resize_mode,
upscaling_resize_w, upscaling_resize_h, upscaling_crop)
cache_key = LruCache.Key(image_hash=hash(np.array(image.getdata()).tobytes()),
cache_key = LruCache.Key(image_hash=image_hash,
info_hash=hash(info),
args_hash=hash((upscale_args, upscale_first)))
args_hash=hash(upscale_args))
cached_entry = cached_images.get(cache_key)
if cached_entry is None:
res = upscale(image, *upscale_args)

View file

@ -22,6 +22,8 @@ from collections import defaultdict, deque
from statistics import stdev, mean
optimizer_dict = {optim_name : cls_obj for optim_name, cls_obj in inspect.getmembers(torch.optim, inspect.isclass) if optim_name != "Optimizer"}
class HypernetworkModule(torch.nn.Module):
multiplier = 1.0
activation_dict = {
@ -35,7 +37,8 @@ class HypernetworkModule(torch.nn.Module):
}
activation_dict.update({cls_name.lower(): cls_obj for cls_name, cls_obj in inspect.getmembers(torch.nn.modules.activation) if inspect.isclass(cls_obj) and cls_obj.__module__ == 'torch.nn.modules.activation'})
def __init__(self, dim, state_dict=None, layer_structure=None, activation_func=None, weight_init='Normal', add_layer_norm=False, use_dropout=False):
def __init__(self, dim, state_dict=None, layer_structure=None, activation_func=None, weight_init='Normal',
add_layer_norm=False, use_dropout=False, activate_output=False, last_layer_dropout=True):
super().__init__()
assert layer_structure is not None, "layer_structure must not be None"
@ -48,8 +51,8 @@ class HypernetworkModule(torch.nn.Module):
# Add a fully-connected layer
linears.append(torch.nn.Linear(int(dim * layer_structure[i]), int(dim * layer_structure[i+1])))
# Add an activation func
if activation_func == "linear" or activation_func is None:
# Add an activation func except last layer
if activation_func == "linear" or activation_func is None or (i >= len(layer_structure) - 2 and not activate_output):
pass
elif activation_func in self.activation_dict:
linears.append(self.activation_dict[activation_func]())
@ -60,8 +63,8 @@ class HypernetworkModule(torch.nn.Module):
if add_layer_norm:
linears.append(torch.nn.LayerNorm(int(dim * layer_structure[i+1])))
# Add dropout expect last layer
if use_dropout and i < len(layer_structure) - 3:
# Add dropout except last layer
if use_dropout and (i < len(layer_structure) - 3 or last_layer_dropout and i < len(layer_structure) - 2):
linears.append(torch.nn.Dropout(p=0.3))
self.linear = torch.nn.Sequential(*linears)
@ -75,7 +78,7 @@ class HypernetworkModule(torch.nn.Module):
w, b = layer.weight.data, layer.bias.data
if weight_init == "Normal" or type(layer) == torch.nn.LayerNorm:
normal_(w, mean=0.0, std=0.01)
normal_(b, mean=0.0, std=0.005)
normal_(b, mean=0.0, std=0)
elif weight_init == 'XavierUniform':
xavier_uniform_(w)
zeros_(b)
@ -127,7 +130,7 @@ class Hypernetwork:
filename = None
name = None
def __init__(self, name=None, enable_sizes=None, layer_structure=None, activation_func=None, weight_init=None, add_layer_norm=False, use_dropout=False):
def __init__(self, name=None, enable_sizes=None, layer_structure=None, activation_func=None, weight_init=None, add_layer_norm=False, use_dropout=False, activate_output=False, **kwargs):
self.filename = None
self.name = name
self.layers = {}
@ -139,11 +142,17 @@ class Hypernetwork:
self.weight_init = weight_init
self.add_layer_norm = add_layer_norm
self.use_dropout = use_dropout
self.activate_output = activate_output
self.last_layer_dropout = kwargs['last_layer_dropout'] if 'last_layer_dropout' in kwargs else True
self.optimizer_name = None
self.optimizer_state_dict = None
for size in enable_sizes or []:
self.layers[size] = (
HypernetworkModule(size, None, self.layer_structure, self.activation_func, self.weight_init, self.add_layer_norm, self.use_dropout),
HypernetworkModule(size, None, self.layer_structure, self.activation_func, self.weight_init, self.add_layer_norm, self.use_dropout),
HypernetworkModule(size, None, self.layer_structure, self.activation_func, self.weight_init,
self.add_layer_norm, self.use_dropout, self.activate_output, last_layer_dropout=self.last_layer_dropout),
HypernetworkModule(size, None, self.layer_structure, self.activation_func, self.weight_init,
self.add_layer_norm, self.use_dropout, self.activate_output, last_layer_dropout=self.last_layer_dropout),
)
def weights(self):
@ -158,6 +167,7 @@ class Hypernetwork:
def save(self, filename):
state_dict = {}
optimizer_saved_dict = {}
for k, v in self.layers.items():
state_dict[k] = (v[0].state_dict(), v[1].state_dict())
@ -171,8 +181,17 @@ class Hypernetwork:
state_dict['use_dropout'] = self.use_dropout
state_dict['sd_checkpoint'] = self.sd_checkpoint
state_dict['sd_checkpoint_name'] = self.sd_checkpoint_name
state_dict['activate_output'] = self.activate_output
state_dict['last_layer_dropout'] = self.last_layer_dropout
if self.optimizer_name is not None:
optimizer_saved_dict['optimizer_name'] = self.optimizer_name
torch.save(state_dict, filename)
if shared.opts.save_optimizer_state and self.optimizer_state_dict:
optimizer_saved_dict['hash'] = sd_models.model_hash(filename)
optimizer_saved_dict['optimizer_state_dict'] = self.optimizer_state_dict
torch.save(optimizer_saved_dict, filename + '.optim')
def load(self, filename):
self.filename = filename
@ -191,12 +210,29 @@ class Hypernetwork:
print(f"Layer norm is set to {self.add_layer_norm}")
self.use_dropout = state_dict.get('use_dropout', False)
print(f"Dropout usage is set to {self.use_dropout}" )
self.activate_output = state_dict.get('activate_output', True)
print(f"Activate last layer is set to {self.activate_output}")
self.last_layer_dropout = state_dict.get('last_layer_dropout', False)
optimizer_saved_dict = torch.load(self.filename + '.optim', map_location = 'cpu') if os.path.exists(self.filename + '.optim') else {}
self.optimizer_name = optimizer_saved_dict.get('optimizer_name', 'AdamW')
print(f"Optimizer name is {self.optimizer_name}")
if sd_models.model_hash(filename) == optimizer_saved_dict.get('hash', None):
self.optimizer_state_dict = optimizer_saved_dict.get('optimizer_state_dict', None)
else:
self.optimizer_state_dict = None
if self.optimizer_state_dict:
print("Loaded existing optimizer from checkpoint")
else:
print("No saved optimizer exists in checkpoint")
for size, sd in state_dict.items():
if type(size) == int:
self.layers[size] = (
HypernetworkModule(size, sd[0], self.layer_structure, self.activation_func, self.weight_init, self.add_layer_norm, self.use_dropout),
HypernetworkModule(size, sd[1], self.layer_structure, self.activation_func, self.weight_init, self.add_layer_norm, self.use_dropout),
HypernetworkModule(size, sd[0], self.layer_structure, self.activation_func, self.weight_init,
self.add_layer_norm, self.use_dropout, self.activate_output, last_layer_dropout=self.last_layer_dropout),
HypernetworkModule(size, sd[1], self.layer_structure, self.activation_func, self.weight_init,
self.add_layer_norm, self.use_dropout, self.activate_output, last_layer_dropout=self.last_layer_dropout),
)
self.name = state_dict.get('name', self.name)
@ -207,11 +243,11 @@ class Hypernetwork:
def list_hypernetworks(path):
res = {}
for filename in glob.iglob(os.path.join(path, '**/*.pt'), recursive=True):
for filename in sorted(glob.iglob(os.path.join(path, '**/*.pt'), recursive=True)):
name = os.path.splitext(os.path.basename(filename))[0]
# Prevent a hypothetical "None.pt" from being listed.
if name != "None":
res[name] = filename
res[name + f"({sd_models.model_hash(filename)})"] = filename
return res
@ -346,6 +382,7 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, data_root, log
shared.state.textinfo = "Initializing hypernetwork training..."
shared.state.job_count = steps
hypernetwork_name = hypernetwork_name.rsplit('(', 1)[0]
filename = os.path.join(shared.cmd_opts.hypernetwork_dir, f'{hypernetwork_name}.pt')
log_directory = os.path.join(log_directory, datetime.datetime.now().strftime("%Y-%m-%d"), hypernetwork_name)
@ -392,8 +429,22 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, data_root, log
weights = hypernetwork.weights()
for weight in weights:
weight.requires_grad = True
# if optimizer == "AdamW": or else Adam / AdamW / SGD, etc...
optimizer = torch.optim.AdamW(weights, lr=scheduler.learn_rate)
# Here we use optimizer from saved HN, or we can specify as UI option.
if hypernetwork.optimizer_name in optimizer_dict:
optimizer = optimizer_dict[hypernetwork.optimizer_name](params=weights, lr=scheduler.learn_rate)
optimizer_name = hypernetwork.optimizer_name
else:
print(f"Optimizer type {hypernetwork.optimizer_name} is not defined!")
optimizer = torch.optim.AdamW(params=weights, lr=scheduler.learn_rate)
optimizer_name = 'AdamW'
if hypernetwork.optimizer_state_dict: # This line must be changed if Optimizer type can be different from saved optimizer.
try:
optimizer.load_state_dict(hypernetwork.optimizer_state_dict)
except RuntimeError as e:
print("Cannot resume from saved optimizer!")
print(e)
steps_without_grad = 0
@ -455,7 +506,11 @@ def train_hypernetwork(hypernetwork_name, learn_rate, batch_size, data_root, log
# Before saving, change name to match current checkpoint.
hypernetwork_name_every = f'{hypernetwork_name}-{steps_done}'
last_saved_file = os.path.join(hypernetwork_dir, f'{hypernetwork_name_every}.pt')
hypernetwork.optimizer_name = optimizer_name
if shared.opts.save_optimizer_state:
hypernetwork.optimizer_state_dict = optimizer.state_dict()
save_hypernetwork(hypernetwork, checkpoint, hypernetwork_name, last_saved_file)
hypernetwork.optimizer_state_dict = None # dereference it after saving, to save memory.
textual_inversion.write_loss(log_directory, "hypernetwork_loss.csv", hypernetwork.step, len(ds), {
"loss": f"{previous_mean_loss:.7f}",
@ -518,8 +573,12 @@ Last saved image: {html.escape(last_saved_image)}<br/>
report_statistics(loss_dict)
filename = os.path.join(shared.cmd_opts.hypernetwork_dir, f'{hypernetwork_name}.pt')
hypernetwork.optimizer_name = optimizer_name
if shared.opts.save_optimizer_state:
hypernetwork.optimizer_state_dict = optimizer.state_dict()
save_hypernetwork(hypernetwork, checkpoint, hypernetwork_name, filename)
del optimizer
hypernetwork.optimizer_state_dict = None # dereference it after saving, to save memory.
return hypernetwork, filename
def save_hypernetwork(hypernetwork, checkpoint, hypernetwork_name, filename):

View file

@ -9,7 +9,7 @@ from modules import devices, sd_hijack, shared
from modules.hypernetworks import hypernetwork
not_available = ["hardswish", "multiheadattention"]
keys = ["linear"] + list(x for x in hypernetwork.HypernetworkModule.activation_dict.keys() if x not in not_available)
keys = list(x for x in hypernetwork.HypernetworkModule.activation_dict.keys() if x not in not_available)
def create_hypernetwork(name, enable_sizes, overwrite_old, layer_structure=None, activation_func=None, weight_init=None, add_layer_norm=False, use_dropout=False):
# Remove illegal characters from name.

View file

@ -81,7 +81,8 @@ def img2img(mode: int, prompt: str, negative_prompt: str, prompt_style: str, pro
mask = None
# Use the EXIF orientation of photos taken by smartphones.
image = ImageOps.exif_transpose(image)
if image is not None:
image = ImageOps.exif_transpose(image)
assert 0. <= denoising_strength <= 1., 'can only work with strength in [0.0, 1.0]'
@ -137,6 +138,8 @@ def img2img(mode: int, prompt: str, negative_prompt: str, prompt_style: str, pro
if processed is None:
processed = process_images(p)
p.close()
shared.total_tqdm.clear()
generation_info_js = processed.js()

View file

@ -101,8 +101,8 @@ class LDSR:
down_sample_rate = target_scale / 4
wd = width_og * down_sample_rate
hd = height_og * down_sample_rate
width_downsampled_pre = int(wd)
height_downsampled_pre = int(hd)
width_downsampled_pre = int(np.ceil(wd))
height_downsampled_pre = int(np.ceil(hd))
if down_sample_rate != 1:
print(
@ -110,7 +110,12 @@ class LDSR:
im_og = im_og.resize((width_downsampled_pre, height_downsampled_pre), Image.LANCZOS)
else:
print(f"Down sample rate is 1 from {target_scale} / 4 (Not downsampling)")
logs = self.run(model["model"], im_og, diffusion_steps, eta)
# pad width and height to multiples of 64, pads with the edge values of image to avoid artifacts
pad_w, pad_h = np.max(((2, 2), np.ceil(np.array(im_og.size) / 64).astype(int)), axis=0) * 64 - im_og.size
im_padded = Image.fromarray(np.pad(np.array(im_og), ((0, pad_h), (0, pad_w), (0, 0)), mode='edge'))
logs = self.run(model["model"], im_padded, diffusion_steps, eta)
sample = logs["sample"]
sample = sample.detach().cpu()
@ -120,6 +125,9 @@ class LDSR:
sample = np.transpose(sample, (0, 2, 3, 1))
a = Image.fromarray(sample[0])
# remove padding
a = a.crop((0, 0) + tuple(np.array(im_og.size) * 4))
del model
gc.collect()
torch.cuda.empty_cache()

View file

@ -3,6 +3,7 @@ import os
import sys
import traceback
localizations = {}
@ -16,6 +17,11 @@ def list_localizations(dirname):
localizations[fn] = os.path.join(dirname, file)
from modules import scripts
for file in scripts.list_scripts("localizations", ".json"):
fn, ext = os.path.splitext(file.filename)
localizations[fn] = file.path
def localization_js(current_localization_name):
fn = localizations.get(current_localization_name, None)

View file

@ -49,7 +49,7 @@ def expand_crop_region(crop_region, processing_width, processing_height, image_w
ratio_processing = processing_width / processing_height
if ratio_crop_region > ratio_processing:
desired_height = (x2 - x1) * ratio_processing
desired_height = (x2 - x1) / ratio_processing
desired_height_diff = int(desired_height - (y2-y1))
y1 -= desired_height_diff//2
y2 += desired_height_diff - desired_height_diff//2

View file

@ -85,6 +85,9 @@ def cleanup_models():
src_path = os.path.join(root_path, "ESRGAN")
dest_path = os.path.join(models_path, "ESRGAN")
move_files(src_path, dest_path)
src_path = os.path.join(models_path, "BSRGAN")
dest_path = os.path.join(models_path, "ESRGAN")
move_files(src_path, dest_path, ".pth")
src_path = os.path.join(root_path, "gfpgan")
dest_path = os.path.join(models_path, "GFPGAN")
move_files(src_path, dest_path)

View file

@ -134,11 +134,7 @@ class StableDiffusionProcessing():
# Dummy zero conditioning if we're not using inpainting model.
# Still takes up a bit of memory, but no encoder call.
# Pretty sure we can just make this a 1x1 image since its not going to be used besides its batch size.
return torch.zeros(
x.shape[0], 5, 1, 1,
dtype=x.dtype,
device=x.device
)
return x.new_zeros(x.shape[0], 5, 1, 1)
height = height or self.height
width = width or self.width
@ -156,11 +152,7 @@ class StableDiffusionProcessing():
def img2img_image_conditioning(self, source_image, latent_image, image_mask = None):
if self.sampler.conditioning_key not in {'hybrid', 'concat'}:
# Dummy zero conditioning if we're not using inpainting model.
return torch.zeros(
latent_image.shape[0], 5, 1, 1,
dtype=latent_image.dtype,
device=latent_image.device
)
return latent_image.new_zeros(latent_image.shape[0], 5, 1, 1)
# Handle the different mask inputs
if image_mask is not None:
@ -174,11 +166,11 @@ class StableDiffusionProcessing():
# Inpainting model uses a discretized mask as input, so we round to either 1.0 or 0.0
conditioning_mask = torch.round(conditioning_mask)
else:
conditioning_mask = torch.ones(1, 1, *source_image.shape[-2:])
conditioning_mask = source_image.new_ones(1, 1, *source_image.shape[-2:])
# Create another latent image, this time with a masked version of the original input.
# Smoothly interpolate between the masked and unmasked latent conditioning image using a parameter.
conditioning_mask = conditioning_mask.to(source_image.device)
conditioning_mask = conditioning_mask.to(source_image.device).to(source_image.dtype)
conditioning_image = torch.lerp(
source_image,
source_image * (1.0 - conditioning_mask),
@ -199,9 +191,13 @@ class StableDiffusionProcessing():
def init(self, all_prompts, all_seeds, all_subseeds):
pass
def sample(self, conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength):
def sample(self, conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength, prompts):
raise NotImplementedError()
def close(self):
self.sd_model = None
self.sampler = None
class Processed:
def __init__(self, p: StableDiffusionProcessing, images_list, seed=-1, info="", subseed=None, all_prompts=None, all_seeds=None, all_subseeds=None, index_of_first_image=0, infotexts=None):
@ -422,13 +418,13 @@ def process_images(p: StableDiffusionProcessing) -> Processed:
try:
for k, v in p.override_settings.items():
opts.data[k] = v # we don't call onchange for simplicity which makes changing model, hypernet impossible
setattr(opts, k, v) # we don't call onchange for simplicity which makes changing model, hypernet impossible
res = process_images_inner(p)
finally:
for k, v in stored_opts.items():
opts.data[k] = v
setattr(opts, k, v)
return res
@ -505,6 +501,9 @@ def process_images_inner(p: StableDiffusionProcessing) -> Processed:
if len(prompts) == 0:
break
if p.scripts is not None:
p.scripts.process_batch(p, batch_number=n, prompts=prompts, seeds=seeds, subseeds=subseeds)
with devices.autocast():
uc = prompt_parser.get_learned_conditioning(shared.sd_model, len(prompts) * [p.negative_prompt], p.steps)
c = prompt_parser.get_multicond_learned_conditioning(shared.sd_model, prompts, p.steps)
@ -517,7 +516,7 @@ def process_images_inner(p: StableDiffusionProcessing) -> Processed:
shared.state.job = f"Batch {n+1} out of {p.n_iter}"
with devices.autocast():
samples_ddim = p.sample(conditioning=c, unconditional_conditioning=uc, seeds=seeds, subseeds=subseeds, subseed_strength=p.subseed_strength)
samples_ddim = p.sample(conditioning=c, unconditional_conditioning=uc, seeds=seeds, subseeds=subseeds, subseed_strength=p.subseed_strength, prompts=prompts)
samples_ddim = samples_ddim.to(devices.dtype_vae)
x_samples_ddim = decode_first_stage(p.sd_model, samples_ddim)
@ -597,9 +596,6 @@ def process_images_inner(p: StableDiffusionProcessing) -> Processed:
if p.scripts is not None:
p.scripts.postprocess(p, res)
p.sd_model = None
p.sampler = None
return res
@ -648,7 +644,7 @@ class StableDiffusionProcessingTxt2Img(StableDiffusionProcessing):
self.truncate_x = int(self.firstphase_width - firstphase_width_truncated) // opt_f
self.truncate_y = int(self.firstphase_height - firstphase_height_truncated) // opt_f
def sample(self, conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength):
def sample(self, conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength, prompts):
self.sampler = sd_samplers.create_sampler_with_index(sd_samplers.samplers, self.sampler_index, self.sd_model)
if not self.enable_hr:
@ -661,9 +657,28 @@ class StableDiffusionProcessingTxt2Img(StableDiffusionProcessing):
samples = samples[:, :, self.truncate_y//2:samples.shape[2]-self.truncate_y//2, self.truncate_x//2:samples.shape[3]-self.truncate_x//2]
"""saves image before applying hires fix, if enabled in options; takes as an arguyment either an image or batch with latent space images"""
def save_intermediate(image, index):
if not opts.save or self.do_not_save_samples or not opts.save_images_before_highres_fix:
return
if not isinstance(image, Image.Image):
image = sd_samplers.sample_to_image(image, index)
images.save_image(image, self.outpath_samples, "", seeds[index], prompts[index], opts.samples_format, suffix="-before-highres-fix")
if opts.use_scale_latent_for_hires_fix:
for i in range(samples.shape[0]):
save_intermediate(samples, i)
samples = torch.nn.functional.interpolate(samples, size=(self.height // opt_f, self.width // opt_f), mode="bilinear")
# Avoid making the inpainting conditioning unless necessary as
# this does need some extra compute to decode / encode the image again.
if getattr(self, "inpainting_mask_weight", shared.opts.inpainting_mask_weight) < 1.0:
image_conditioning = self.img2img_image_conditioning(decode_first_stage(self.sd_model, samples), samples)
else:
image_conditioning = self.txt2img_image_conditioning(samples)
else:
decoded_samples = decode_first_stage(self.sd_model, samples)
lowres_samples = torch.clamp((decoded_samples + 1.0) / 2.0, min=0.0, max=1.0)
@ -673,6 +688,9 @@ class StableDiffusionProcessingTxt2Img(StableDiffusionProcessing):
x_sample = 255. * np.moveaxis(x_sample.cpu().numpy(), 0, 2)
x_sample = x_sample.astype(np.uint8)
image = Image.fromarray(x_sample)
save_intermediate(image, i)
image = images.resize_image(0, image, self.width, self.height)
image = np.array(image).astype(np.float32) / 255.0
image = np.moveaxis(image, 2, 0)
@ -684,14 +702,14 @@ class StableDiffusionProcessingTxt2Img(StableDiffusionProcessing):
samples = self.sd_model.get_first_stage_encoding(self.sd_model.encode_first_stage(decoded_samples))
image_conditioning = self.img2img_image_conditioning(decoded_samples, samples)
shared.state.nextjob()
self.sampler = sd_samplers.create_sampler_with_index(sd_samplers.samplers, self.sampler_index, self.sd_model)
noise = create_random_tensors(samples.shape[1:], seeds=seeds, subseeds=subseeds, subseed_strength=subseed_strength, seed_resize_from_h=self.seed_resize_from_h, seed_resize_from_w=self.seed_resize_from_w, p=self)
image_conditioning = self.txt2img_image_conditioning(x)
# GC now before running the next img2img to prevent running out of memory
x = None
devices.torch_gc()
@ -830,8 +848,7 @@ class StableDiffusionProcessingImg2Img(StableDiffusionProcessing):
self.image_conditioning = self.img2img_image_conditioning(image, self.init_latent, self.image_mask)
def sample(self, conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength):
def sample(self, conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength, prompts):
x = create_random_tensors([opt_C, self.height // opt_f, self.width // opt_f], seeds=seeds, subseeds=subseeds, subseed_strength=self.subseed_strength, seed_resize_from_h=self.seed_resize_from_h, seed_resize_from_w=self.seed_resize_from_w, p=self)
samples = self.sampler.sample_img2img(self, self.init_latent, x, conditioning, unconditional_conditioning, image_conditioning=self.image_conditioning)
@ -842,4 +859,4 @@ class StableDiffusionProcessingImg2Img(StableDiffusionProcessing):
del x
devices.torch_gc()
return samples
return samples

View file

@ -23,11 +23,18 @@ def encode(*args):
class RestrictedUnpickler(pickle.Unpickler):
extra_handler = None
def persistent_load(self, saved_id):
assert saved_id[0] == 'storage'
return TypedStorage()
def find_class(self, module, name):
if self.extra_handler is not None:
res = self.extra_handler(module, name)
if res is not None:
return res
if module == 'collections' and name == 'OrderedDict':
return getattr(collections, name)
if module == 'torch._utils' and name in ['_rebuild_tensor_v2', '_rebuild_parameter']:
@ -52,7 +59,7 @@ class RestrictedUnpickler(pickle.Unpickler):
return set
# Forbid everything else.
raise pickle.UnpicklingError(f"global '{module}/{name}' is forbidden")
raise Exception(f"global '{module}/{name}' is forbidden")
allowed_zip_names = ["archive/data.pkl", "archive/version"]
@ -69,7 +76,7 @@ def check_zip_filenames(filename, names):
raise Exception(f"bad file inside {filename}: {name}")
def check_pt(filename):
def check_pt(filename, extra_handler):
try:
# new pytorch format is a zip file
@ -78,6 +85,7 @@ def check_pt(filename):
with z.open('archive/data.pkl') as file:
unpickler = RestrictedUnpickler(file)
unpickler.extra_handler = extra_handler
unpickler.load()
except zipfile.BadZipfile:
@ -85,16 +93,42 @@ def check_pt(filename):
# if it's not a zip file, it's an olf pytorch format, with five objects written to pickle
with open(filename, "rb") as file:
unpickler = RestrictedUnpickler(file)
unpickler.extra_handler = extra_handler
for i in range(5):
unpickler.load()
def load(filename, *args, **kwargs):
return load_with_extra(filename, *args, **kwargs)
def load_with_extra(filename, extra_handler=None, *args, **kwargs):
"""
this functon is intended to be used by extensions that want to load models with
some extra classes in them that the usual unpickler would find suspicious.
Use the extra_handler argument to specify a function that takes module and field name as text,
and returns that field's value:
```python
def extra(module, name):
if module == 'collections' and name == 'OrderedDict':
return collections.OrderedDict
return None
safe.load_with_extra('model.pt', extra_handler=extra)
```
The alternative to this is just to use safe.unsafe_torch_load('model.pt'), which as the name implies is
definitely unsafe.
"""
from modules import shared
try:
if not shared.cmd_opts.disable_safe_unpickle:
check_pt(filename)
check_pt(filename, extra_handler)
except pickle.UnpicklingError:
print(f"Error verifying pickled file from {filename}:", file=sys.stderr)

View file

@ -2,6 +2,7 @@ import sys
import traceback
from collections import namedtuple
import inspect
from typing import Optional
from fastapi import FastAPI
from gradio import Blocks
@ -26,25 +27,42 @@ class ImageSaveParams:
"""dictionary with parameters for image's PNG info data; infotext will have the key 'parameters'"""
class CFGDenoiserParams:
def __init__(self, x, image_cond, sigma, sampling_step, total_sampling_steps):
self.x = x
"""Latent image representation in the process of being denoised"""
self.image_cond = image_cond
"""Conditioning image"""
self.sigma = sigma
"""Current sigma noise step value"""
self.sampling_step = sampling_step
"""Current Sampling step number"""
self.total_sampling_steps = total_sampling_steps
"""Total number of sampling steps planned"""
ScriptCallback = namedtuple("ScriptCallback", ["script", "callback"])
callbacks_app_started = []
callbacks_model_loaded = []
callbacks_ui_tabs = []
callbacks_ui_settings = []
callbacks_before_image_saved = []
callbacks_image_saved = []
callback_map = dict(
callbacks_app_started=[],
callbacks_model_loaded=[],
callbacks_ui_tabs=[],
callbacks_ui_settings=[],
callbacks_before_image_saved=[],
callbacks_image_saved=[],
callbacks_cfg_denoiser=[]
)
def clear_callbacks():
callbacks_model_loaded.clear()
callbacks_ui_tabs.clear()
callbacks_ui_settings.clear()
callbacks_before_image_saved.clear()
callbacks_image_saved.clear()
for callback_list in callback_map.values():
callback_list.clear()
def app_started_callback(demo: Blocks, app: FastAPI):
for c in callbacks_app_started:
def app_started_callback(demo: Optional[Blocks], app: FastAPI):
for c in callback_map['callbacks_app_started']:
try:
c.callback(demo, app)
except Exception:
@ -52,7 +70,7 @@ def app_started_callback(demo: Blocks, app: FastAPI):
def model_loaded_callback(sd_model):
for c in callbacks_model_loaded:
for c in callback_map['callbacks_model_loaded']:
try:
c.callback(sd_model)
except Exception:
@ -62,7 +80,7 @@ def model_loaded_callback(sd_model):
def ui_tabs_callback():
res = []
for c in callbacks_ui_tabs:
for c in callback_map['callbacks_ui_tabs']:
try:
res += c.callback() or []
except Exception:
@ -72,7 +90,7 @@ def ui_tabs_callback():
def ui_settings_callback():
for c in callbacks_ui_settings:
for c in callback_map['callbacks_ui_settings']:
try:
c.callback()
except Exception:
@ -80,7 +98,7 @@ def ui_settings_callback():
def before_image_saved_callback(params: ImageSaveParams):
for c in callbacks_before_image_saved:
for c in callback_map['callbacks_before_image_saved']:
try:
c.callback(params)
except Exception:
@ -88,30 +106,54 @@ def before_image_saved_callback(params: ImageSaveParams):
def image_saved_callback(params: ImageSaveParams):
for c in callbacks_image_saved:
for c in callback_map['callbacks_image_saved']:
try:
c.callback(params)
except Exception:
report_exception(c, 'image_saved_callback')
def cfg_denoiser_callback(params: CFGDenoiserParams):
for c in callback_map['callbacks_cfg_denoiser']:
try:
c.callback(params)
except Exception:
report_exception(c, 'cfg_denoiser_callback')
def add_callback(callbacks, fun):
stack = [x for x in inspect.stack() if x.filename != __file__]
filename = stack[0].filename if len(stack) > 0 else 'unknown file'
callbacks.append(ScriptCallback(filename, fun))
def remove_current_script_callbacks():
stack = [x for x in inspect.stack() if x.filename != __file__]
filename = stack[0].filename if len(stack) > 0 else 'unknown file'
if filename == 'unknown file':
return
for callback_list in callback_map.values():
for callback_to_remove in [cb for cb in callback_list if cb.script == filename]:
callback_list.remove(callback_to_remove)
def remove_callbacks_for_function(callback_func):
for callback_list in callback_map.values():
for callback_to_remove in [cb for cb in callback_list if cb.callback == callback_func]:
callback_list.remove(callback_to_remove)
def on_app_started(callback):
"""register a function to be called when the webui started, the gradio `Block` component and
fastapi `FastAPI` object are passed as the arguments"""
add_callback(callbacks_app_started, callback)
add_callback(callback_map['callbacks_app_started'], callback)
def on_model_loaded(callback):
"""register a function to be called when the stable diffusion model is created; the model is
passed as an argument"""
add_callback(callbacks_model_loaded, callback)
add_callback(callback_map['callbacks_model_loaded'], callback)
def on_ui_tabs(callback):
@ -124,13 +166,13 @@ def on_ui_tabs(callback):
title is tab text displayed to user in the UI
elem_id is HTML id for the tab
"""
add_callback(callbacks_ui_tabs, callback)
add_callback(callback_map['callbacks_ui_tabs'], callback)
def on_ui_settings(callback):
"""register a function to be called before UI settings are populated; add your settings
by using shared.opts.add_option(shared.OptionInfo(...)) """
add_callback(callbacks_ui_settings, callback)
add_callback(callback_map['callbacks_ui_settings'], callback)
def on_before_image_saved(callback):
@ -138,7 +180,7 @@ def on_before_image_saved(callback):
The callback is called with one argument:
- params: ImageSaveParams - parameters the image is to be saved with. You can change fields in this object.
"""
add_callback(callbacks_before_image_saved, callback)
add_callback(callback_map['callbacks_before_image_saved'], callback)
def on_image_saved(callback):
@ -146,4 +188,12 @@ def on_image_saved(callback):
The callback is called with one argument:
- params: ImageSaveParams - parameters the image was saved with. Changing fields in this object does nothing.
"""
add_callback(callbacks_image_saved, callback)
add_callback(callback_map['callbacks_image_saved'], callback)
def on_cfg_denoiser(callback):
"""register a function to be called in the kdiffussion cfg_denoiser method after building the inner model inputs.
The callback is called with one argument:
- params: CFGDenoiserParams - parameters to be passed to the inner model and sampling state details.
"""
add_callback(callback_map['callbacks_cfg_denoiser'], callback)

View file

@ -3,7 +3,6 @@ import sys
import traceback
from collections import namedtuple
import modules.ui as ui
import gradio as gr
from modules.processing import StableDiffusionProcessing
@ -18,6 +17,9 @@ class Script:
args_to = None
alwayson = False
"""A gr.Group component that has all script's UI inside it"""
group = None
infotext_fields = None
"""if set in ui(), this is a list of pairs of gradio component + text; the text will be used when
parsing infotext to set the value for the component; see ui.py's txt2img_paste_fields for an example
@ -70,6 +72,19 @@ class Script:
pass
def process_batch(self, p, *args, **kwargs):
"""
Same as process(), but called for every batch.
**kwargs will have those items:
- batch_number - index of current batch, from 0 to number of batches-1
- prompts - list of prompts for current batch; you can change contents of this list but changing the number of entries will likely break things
- seeds - list of seeds for current batch
- subseeds - list of subseeds for current batch
"""
pass
def postprocess(self, p, processed, *args):
"""
This function is called after processing ends for AlwaysVisible scripts.
@ -218,8 +233,6 @@ class ScriptRunner:
for control in controls:
control.custom_script_source = os.path.basename(script.filename)
if not script.alwayson:
control.visible = False
if script.infotext_fields is not None:
self.infotext_fields += script.infotext_fields
@ -229,40 +242,41 @@ class ScriptRunner:
script.args_to = len(inputs)
for script in self.alwayson_scripts:
with gr.Group():
with gr.Group() as group:
create_script_ui(script, inputs, inputs_alwayson)
script.group = group
dropdown = gr.Dropdown(label="Script", elem_id="script_list", choices=["None"] + self.titles, value="None", type="index")
dropdown.save_to_config = True
inputs[0] = dropdown
for script in self.selectable_scripts:
create_script_ui(script, inputs, inputs_alwayson)
with gr.Group(visible=False) as group:
create_script_ui(script, inputs, inputs_alwayson)
script.group = group
def select_script(script_index):
if 0 < script_index <= len(self.selectable_scripts):
script = self.selectable_scripts[script_index-1]
args_from = script.args_from
args_to = script.args_to
else:
args_from = 0
args_to = 0
selected_script = self.selectable_scripts[script_index - 1] if script_index>0 else None
return [ui.gr_show(True if i == 0 else args_from <= i < args_to or is_alwayson) for i, is_alwayson in enumerate(inputs_alwayson)]
return [gr.update(visible=selected_script == s) for s in self.selectable_scripts]
def init_field(title):
"""called when an initial value is set from ui-config.json to show script's UI components"""
if title == 'None':
return
script_index = self.titles.index(title)
script = self.selectable_scripts[script_index]
for i in range(script.args_from, script.args_to):
inputs[i].visible = True
self.selectable_scripts[script_index].group.visible = True
dropdown.init_field = init_field
dropdown.change(
fn=select_script,
inputs=[dropdown],
outputs=inputs
outputs=[script.group for script in self.selectable_scripts]
)
return inputs
@ -294,6 +308,15 @@ class ScriptRunner:
print(f"Error running process: {script.filename}", file=sys.stderr)
print(traceback.format_exc(), file=sys.stderr)
def process_batch(self, p, **kwargs):
for script in self.alwayson_scripts:
try:
script_args = p.script_args[script.args_from:script.args_to]
script.process_batch(p, *script_args, **kwargs)
except Exception:
print(f"Error running process_batch: {script.filename}", file=sys.stderr)
print(traceback.format_exc(), file=sys.stderr)
def postprocess(self, p, processed):
for script in self.alwayson_scripts:
try:

View file

@ -9,7 +9,7 @@ from omegaconf import OmegaConf
from ldm.util import instantiate_from_config
from modules import shared, modelloader, devices, script_callbacks
from modules import shared, modelloader, devices, script_callbacks, sd_vae
from modules.paths import models_path
from modules.sd_hijack_inpainting import do_inpainting_hijack, should_hijack_inpainting
@ -159,13 +159,16 @@ def get_state_dict_from_checkpoint(pl_sd):
return pl_sd
vae_ignore_keys = {"model_ema.decay", "model_ema.num_updates"}
def load_model_weights(model, checkpoint_info):
def load_model_weights(model, checkpoint_info, vae_file="auto"):
checkpoint_file = checkpoint_info.filename
sd_model_hash = checkpoint_info.hash
if shared.opts.sd_checkpoint_cache > 0 and hasattr(model, "sd_checkpoint_info"):
sd_vae.restore_base_vae(model)
checkpoints_loaded[model.sd_checkpoint_info] = model.state_dict().copy()
vae_file = sd_vae.resolve_vae(checkpoint_file, vae_file=vae_file)
if checkpoint_info not in checkpoints_loaded:
print(f"Loading weights [{sd_model_hash}] from {checkpoint_file}")
@ -182,37 +185,36 @@ def load_model_weights(model, checkpoint_info):
model.to(memory_format=torch.channels_last)
if not shared.cmd_opts.no_half:
vae = model.first_stage_model
# with --no-half-vae, remove VAE from model when doing half() to prevent its weights from being converted to float16
if shared.cmd_opts.no_half_vae:
model.first_stage_model = None
model.half()
model.first_stage_model = vae
devices.dtype = torch.float32 if shared.cmd_opts.no_half else torch.float16
devices.dtype_vae = torch.float32 if shared.cmd_opts.no_half or shared.cmd_opts.no_half_vae else torch.float16
vae_file = os.path.splitext(checkpoint_file)[0] + ".vae.pt"
if not os.path.exists(vae_file) and shared.cmd_opts.vae_path is not None:
vae_file = shared.cmd_opts.vae_path
if os.path.exists(vae_file):
print(f"Loading VAE weights from: {vae_file}")
vae_ckpt = torch.load(vae_file, map_location=shared.weight_load_location)
vae_dict = {k: v for k, v in vae_ckpt["state_dict"].items() if k[0:4] != "loss" and k not in vae_ignore_keys}
model.first_stage_model.load_state_dict(vae_dict)
model.first_stage_model.to(devices.dtype_vae)
if shared.opts.sd_checkpoint_cache > 0:
checkpoints_loaded[checkpoint_info] = model.state_dict().copy()
while len(checkpoints_loaded) > shared.opts.sd_checkpoint_cache:
checkpoints_loaded.popitem(last=False) # LRU
else:
print(f"Loading weights [{sd_model_hash}] from cache")
checkpoints_loaded.move_to_end(checkpoint_info)
vae_name = sd_vae.get_filename(vae_file) if vae_file else None
vae_message = f" with {vae_name} VAE" if vae_name else ""
print(f"Loading weights [{sd_model_hash}]{vae_message} from cache")
model.load_state_dict(checkpoints_loaded[checkpoint_info])
if shared.opts.sd_checkpoint_cache > 0:
while len(checkpoints_loaded) > shared.opts.sd_checkpoint_cache:
checkpoints_loaded.popitem(last=False) # LRU
model.sd_model_hash = sd_model_hash
model.sd_model_checkpoint = checkpoint_file
model.sd_checkpoint_info = checkpoint_info
sd_vae.load_vae(model, vae_file)
def load_model(checkpoint_info=None):
from modules import lowvram, sd_hijack
@ -263,7 +265,7 @@ def load_model(checkpoint_info=None):
def reload_model_weights(sd_model=None, info=None):
from modules import lowvram, devices, sd_hijack
checkpoint_info = info or select_checkpoint()
if not sd_model:
sd_model = shared.sd_model

View file

@ -12,6 +12,7 @@ from modules import prompt_parser, devices, processing, images
from modules.shared import opts, cmd_opts, state
import modules.shared as shared
from modules.script_callbacks import CFGDenoiserParams, cfg_denoiser_callback
SamplerData = namedtuple('SamplerData', ['name', 'constructor', 'aliases', 'options'])
@ -23,11 +24,15 @@ samplers_k_diffusion = [
('Heun', 'sample_heun', ['k_heun'], {}),
('DPM2', 'sample_dpm_2', ['k_dpm_2'], {}),
('DPM2 a', 'sample_dpm_2_ancestral', ['k_dpm_2_a'], {}),
('DPM++ 2S a', 'sample_dpmpp_2s_ancestral', ['k_dpmpp_2s_a'], {}),
('DPM++ 2M', 'sample_dpmpp_2m', ['k_dpmpp_2m'], {}),
('DPM fast', 'sample_dpm_fast', ['k_dpm_fast'], {}),
('DPM adaptive', 'sample_dpm_adaptive', ['k_dpm_ad'], {}),
('LMS Karras', 'sample_lms', ['k_lms_ka'], {'scheduler': 'karras'}),
('DPM2 Karras', 'sample_dpm_2', ['k_dpm_2_ka'], {'scheduler': 'karras'}),
('DPM2 a Karras', 'sample_dpm_2_ancestral', ['k_dpm_2_a_ka'], {'scheduler': 'karras'}),
('DPM++ 2S a Karras', 'sample_dpmpp_2s_ancestral', ['k_dpmpp_2s_a_ka'], {'scheduler': 'karras'}),
('DPM++ 2M Karras', 'sample_dpmpp_2m', ['k_dpmpp_2m_ka'], {'scheduler': 'karras'}),
]
samplers_data_k_diffusion = [
@ -92,8 +97,8 @@ def single_sample_to_image(sample):
return Image.fromarray(x_sample)
def sample_to_image(samples):
return single_sample_to_image(samples[0])
def sample_to_image(samples, index=0):
return single_sample_to_image(samples[index])
def samples_to_image_grid(samples):
@ -280,6 +285,12 @@ class CFGDenoiser(torch.nn.Module):
image_cond_in = torch.cat([torch.stack([image_cond[i] for _ in range(n)]) for i, n in enumerate(repeats)] + [image_cond])
sigma_in = torch.cat([torch.stack([sigma[i] for _ in range(n)]) for i, n in enumerate(repeats)] + [sigma])
denoiser_params = CFGDenoiserParams(x_in, image_cond_in, sigma_in, state.sampling_step, state.sampling_steps)
cfg_denoiser_callback(denoiser_params)
x_in = denoiser_params.x
image_cond_in = denoiser_params.image_cond
sigma_in = denoiser_params.sigma
if tensor.shape[1] == uncond.shape[1]:
cond_in = torch.cat([tensor, uncond])

207
modules/sd_vae.py Normal file
View file

@ -0,0 +1,207 @@
import torch
import os
from collections import namedtuple
from modules import shared, devices, script_callbacks
from modules.paths import models_path
import glob
model_dir = "Stable-diffusion"
model_path = os.path.abspath(os.path.join(models_path, model_dir))
vae_dir = "VAE"
vae_path = os.path.abspath(os.path.join(models_path, vae_dir))
vae_ignore_keys = {"model_ema.decay", "model_ema.num_updates"}
default_vae_dict = {"auto": "auto", "None": "None"}
default_vae_list = ["auto", "None"]
default_vae_values = [default_vae_dict[x] for x in default_vae_list]
vae_dict = dict(default_vae_dict)
vae_list = list(default_vae_list)
first_load = True
base_vae = None
loaded_vae_file = None
checkpoint_info = None
def get_base_vae(model):
if base_vae is not None and checkpoint_info == model.sd_checkpoint_info and model:
return base_vae
return None
def store_base_vae(model):
global base_vae, checkpoint_info
if checkpoint_info != model.sd_checkpoint_info:
base_vae = model.first_stage_model.state_dict().copy()
checkpoint_info = model.sd_checkpoint_info
def delete_base_vae():
global base_vae, checkpoint_info
base_vae = None
checkpoint_info = None
def restore_base_vae(model):
global base_vae, checkpoint_info
if base_vae is not None and checkpoint_info == model.sd_checkpoint_info:
load_vae_dict(model, base_vae)
delete_base_vae()
def get_filename(filepath):
return os.path.splitext(os.path.basename(filepath))[0]
def refresh_vae_list(vae_path=vae_path, model_path=model_path):
global vae_dict, vae_list
res = {}
candidates = [
*glob.iglob(os.path.join(model_path, '**/*.vae.ckpt'), recursive=True),
*glob.iglob(os.path.join(model_path, '**/*.vae.pt'), recursive=True),
*glob.iglob(os.path.join(vae_path, '**/*.ckpt'), recursive=True),
*glob.iglob(os.path.join(vae_path, '**/*.pt'), recursive=True)
]
if shared.cmd_opts.vae_path is not None and os.path.isfile(shared.cmd_opts.vae_path):
candidates.append(shared.cmd_opts.vae_path)
for filepath in candidates:
name = get_filename(filepath)
res[name] = filepath
vae_list.clear()
vae_list.extend(default_vae_list)
vae_list.extend(list(res.keys()))
vae_dict.clear()
vae_dict.update(res)
vae_dict.update(default_vae_dict)
return vae_list
def resolve_vae(checkpoint_file, vae_file="auto"):
global first_load, vae_dict, vae_list
# if vae_file argument is provided, it takes priority, but not saved
if vae_file and vae_file not in default_vae_list:
if not os.path.isfile(vae_file):
vae_file = "auto"
print("VAE provided as function argument doesn't exist")
# for the first load, if vae-path is provided, it takes priority, saved, and failure is reported
if first_load and shared.cmd_opts.vae_path is not None:
if os.path.isfile(shared.cmd_opts.vae_path):
vae_file = shared.cmd_opts.vae_path
shared.opts.data['sd_vae'] = get_filename(vae_file)
else:
print("VAE provided as command line argument doesn't exist")
# else, we load from settings
if vae_file == "auto" and shared.opts.sd_vae is not None:
# if saved VAE settings isn't recognized, fallback to auto
vae_file = vae_dict.get(shared.opts.sd_vae, "auto")
# if VAE selected but not found, fallback to auto
if vae_file not in default_vae_values and not os.path.isfile(vae_file):
vae_file = "auto"
print("Selected VAE doesn't exist")
# vae-path cmd arg takes priority for auto
if vae_file == "auto" and shared.cmd_opts.vae_path is not None:
if os.path.isfile(shared.cmd_opts.vae_path):
vae_file = shared.cmd_opts.vae_path
print("Using VAE provided as command line argument")
# if still not found, try look for ".vae.pt" beside model
model_path = os.path.splitext(checkpoint_file)[0]
if vae_file == "auto":
vae_file_try = model_path + ".vae.pt"
if os.path.isfile(vae_file_try):
vae_file = vae_file_try
print("Using VAE found beside selected model")
# if still not found, try look for ".vae.ckpt" beside model
if vae_file == "auto":
vae_file_try = model_path + ".vae.ckpt"
if os.path.isfile(vae_file_try):
vae_file = vae_file_try
print("Using VAE found beside selected model")
# No more fallbacks for auto
if vae_file == "auto":
vae_file = None
# Last check, just because
if vae_file and not os.path.exists(vae_file):
vae_file = None
return vae_file
def load_vae(model, vae_file=None):
global first_load, vae_dict, vae_list, loaded_vae_file
# save_settings = False
if vae_file:
print(f"Loading VAE weights from: {vae_file}")
vae_ckpt = torch.load(vae_file, map_location=shared.weight_load_location)
vae_dict_1 = {k: v for k, v in vae_ckpt["state_dict"].items() if k[0:4] != "loss" and k not in vae_ignore_keys}
load_vae_dict(model, vae_dict_1)
# If vae used is not in dict, update it
# It will be removed on refresh though
vae_opt = get_filename(vae_file)
if vae_opt not in vae_dict:
vae_dict[vae_opt] = vae_file
vae_list.append(vae_opt)
loaded_vae_file = vae_file
"""
# Save current VAE to VAE settings, maybe? will it work?
if save_settings:
if vae_file is None:
vae_opt = "None"
# shared.opts.sd_vae = vae_opt
"""
first_load = False
# don't call this from outside
def load_vae_dict(model, vae_dict_1=None):
if vae_dict_1:
store_base_vae(model)
model.first_stage_model.load_state_dict(vae_dict_1)
else:
restore_base_vae()
model.first_stage_model.to(devices.dtype_vae)
def reload_vae_weights(sd_model=None, vae_file="auto"):
from modules import lowvram, devices, sd_hijack
if not sd_model:
sd_model = shared.sd_model
checkpoint_info = sd_model.sd_checkpoint_info
checkpoint_file = checkpoint_info.filename
vae_file = resolve_vae(checkpoint_file, vae_file=vae_file)
if loaded_vae_file == vae_file:
return
if shared.cmd_opts.lowvram or shared.cmd_opts.medvram:
lowvram.send_everything_to_cpu()
else:
sd_model.to(devices.cpu)
sd_hijack.model_hijack.undo_hijack(sd_model)
load_vae(sd_model, vae_file)
sd_hijack.model_hijack.hijack(sd_model)
script_callbacks.model_loaded_callback(sd_model)
if not shared.cmd_opts.lowvram and not shared.cmd_opts.medvram:
sd_model.to(devices.device)
print(f"VAE Weights loaded.")
return sd_model

View file

@ -4,6 +4,7 @@ import json
import os
import sys
from collections import OrderedDict
import time
import gradio as gr
import tqdm
@ -14,7 +15,7 @@ import modules.memmon
import modules.sd_models
import modules.styles
import modules.devices as devices
from modules import sd_samplers, sd_models, localization
from modules import sd_samplers, sd_models, localization, sd_vae
from modules.hypernetworks import hypernetwork
from modules.paths import models_path, script_path, sd_path
@ -43,6 +44,7 @@ parser.add_argument("--precision", type=str, help="evaluate at this precision",
parser.add_argument("--share", action='store_true', help="use share=True for gradio and make the UI accessible through their site")
parser.add_argument("--ngrok", type=str, help="ngrok authtoken, alternative to gradio --share", default=None)
parser.add_argument("--ngrok-region", type=str, help="The region in which ngrok should start.", default="us")
parser.add_argument("--enable-insecure-extension-access", action='store_true', help="enable extensions tab regardless of other options")
parser.add_argument("--codeformer-models-path", type=str, help="Path to directory with codeformer model file(s).", default=os.path.join(models_path, 'Codeformer'))
parser.add_argument("--gfpgan-models-path", type=str, help="Path to directory with GFPGAN model file(s).", default=os.path.join(models_path, 'GFPGAN'))
parser.add_argument("--esrgan-models-path", type=str, help="Path to directory with ESRGAN model file(s).", default=os.path.join(models_path, 'ESRGAN'))
@ -84,6 +86,10 @@ parser.add_argument("--nowebui", action='store_true', help="use api=True to laun
parser.add_argument("--ui-debug-mode", action='store_true', help="Don't load model to quickly launch UI")
parser.add_argument("--device-id", type=str, help="Select the default CUDA device to use (export CUDA_VISIBLE_DEVICES=0,1,etc might be needed before)", default=None)
parser.add_argument("--administrator", action='store_true', help="Administrator rights", default=False)
parser.add_argument("--cors-allow-origins", type=str, help="Allowed CORS origins", default=None)
parser.add_argument("--tls-keyfile", type=str, help="Partially enables TLS, requires --tls-certfile to fully function", default=None)
parser.add_argument("--tls-certfile", type=str, help="Partially enables TLS, requires --tls-keyfile to fully function", default=None)
parser.add_argument("--server-name", type=str, help="Sets hostname of server", default=None)
cmd_opts = parser.parse_args()
restricted_opts = {
@ -98,7 +104,7 @@ restricted_opts = {
"outdir_save",
}
cmd_opts.disable_extension_access = cmd_opts.share or cmd_opts.listen
cmd_opts.disable_extension_access = (cmd_opts.share or cmd_opts.listen) and not cmd_opts.enable_insecure_extension_access
devices.device, devices.device_interrogate, devices.device_gfpgan, devices.device_swinir, devices.device_esrgan, devices.device_scunet, devices.device_codeformer = \
(devices.cpu if any(y in cmd_opts.use_cpu for y in [x, 'all']) else devices.get_optimal_device() for x in ['sd', 'interrogate', 'gfpgan', 'swinir', 'esrgan', 'scunet', 'codeformer'])
@ -135,6 +141,7 @@ class State:
current_image = None
current_image_sampling_step = 0
textinfo = None
time_start = None
need_restart = False
def skip(self):
@ -144,6 +151,9 @@ class State:
self.interrupted = True
def nextjob(self):
if opts.show_progress_every_n_steps == -1:
self.do_set_current_image()
self.job_no += 1
self.sampling_step = 0
self.current_image_sampling_step = 0
@ -172,6 +182,7 @@ class State:
self.skipped = False
self.interrupted = False
self.textinfo = None
self.time_start = time.time()
devices.torch_gc()
@ -181,6 +192,24 @@ class State:
devices.torch_gc()
"""sets self.current_image from self.current_latent if enough sampling steps have been made after the last call to this"""
def set_current_image(self):
if self.sampling_step - self.current_image_sampling_step >= opts.show_progress_every_n_steps and opts.show_progress_every_n_steps > 0:
self.do_set_current_image()
def do_set_current_image(self):
if not parallel_processing_allowed:
return
if self.current_latent is None:
return
if opts.show_progress_grid:
self.current_image = sd_samplers.samples_to_image_grid(self.current_latent)
else:
self.current_image = sd_samplers.sample_to_image(self.current_latent)
self.current_image_sampling_step = self.sampling_step
state = State()
artist_db = modules.artists.ArtistsDatabase(os.path.join(script_path, 'artists.csv'))
@ -192,8 +221,6 @@ interrogator = modules.interrogate.InterrogateModels("interrogate")
face_restorers = []
localization.list_localizations(cmd_opts.localizations_dir)
def realesrgan_models_names():
import modules.realesrgan_model
@ -238,6 +265,8 @@ options_templates.update(options_section(('saving-images', "Saving images/grids"
"enable_pnginfo": OptionInfo(True, "Save text information about generation parameters as chunks to png files"),
"save_txt": OptionInfo(False, "Create a text file next to every image with generation parameters."),
"save_images_before_face_restoration": OptionInfo(False, "Save a copy of image before doing face restoration."),
"save_images_before_highres_fix": OptionInfo(False, "Save a copy of image before applying highres fix."),
"save_images_before_color_correction": OptionInfo(False, "Save a copy of image before applying color correction to img2img results"),
"jpeg_quality": OptionInfo(80, "Quality for saved jpeg images", gr.Slider, {"minimum": 1, "maximum": 100, "step": 1}),
"export_for_4chan": OptionInfo(True, "If PNG image is larger than 4MB or any dimension is larger than 4000, downscale and save copy as JPG"),
@ -290,6 +319,7 @@ options_templates.update(options_section(('system', "System"), {
options_templates.update(options_section(('training', "Training"), {
"unload_models_when_training": OptionInfo(False, "Move VAE and CLIP to RAM when training if possible. Saves VRAM."),
"save_optimizer_state": OptionInfo(False, "Saves Optimizer state as separate *.optim file. Training can be resumed with HN itself and matching optim file."),
"dataset_filename_word_regex": OptionInfo("", "Filename word regex"),
"dataset_filename_join_string": OptionInfo(" ", "Filename join string"),
"training_image_repeats_per_epoch": OptionInfo(1, "Number of repeats for a single input image per epoch; used only for displaying epoch number", gr.Number, {"precision": 0}),
@ -300,11 +330,11 @@ options_templates.update(options_section(('training', "Training"), {
options_templates.update(options_section(('sd', "Stable Diffusion"), {
"sd_model_checkpoint": OptionInfo(None, "Stable Diffusion checkpoint", gr.Dropdown, lambda: {"choices": modules.sd_models.checkpoint_tiles()}, refresh=sd_models.list_models),
"sd_checkpoint_cache": OptionInfo(0, "Checkpoints to cache in RAM", gr.Slider, {"minimum": 0, "maximum": 10, "step": 1}),
"sd_vae": OptionInfo("auto", "SD VAE", gr.Dropdown, lambda: {"choices": list(sd_vae.vae_list)}, refresh=sd_vae.refresh_vae_list),
"sd_hypernetwork": OptionInfo("None", "Hypernetwork", gr.Dropdown, lambda: {"choices": ["None"] + [x for x in hypernetworks.keys()]}, refresh=reload_hypernetworks),
"sd_hypernetwork_strength": OptionInfo(1.0, "Hypernetwork strength", gr.Slider, {"minimum": 0.0, "maximum": 1.0, "step": 0.001}),
"inpainting_mask_weight": OptionInfo(1.0, "Inpainting conditioning mask strength", gr.Slider, {"minimum": 0.0, "maximum": 1.0, "step": 0.01}),
"img2img_color_correction": OptionInfo(False, "Apply color correction to img2img results to match original colors."),
"save_images_before_color_correction": OptionInfo(False, "Save a copy of image before applying color correction to img2img results"),
"img2img_fix_steps": OptionInfo(False, "With img2img, do exactly the amount of steps the slider specifies (normally you'd do less with less denoising)."),
"enable_quantization": OptionInfo(False, "Enable quantization in K samplers for sharper and cleaner results. This may change existing seeds. Requires restart to apply."),
"enable_emphasis": OptionInfo(True, "Emphasis: use (text) to make model pay more attention to text and [text] to make it pay less attention"),
@ -332,7 +362,7 @@ options_templates.update(options_section(('interrogate', "Interrogate Options"),
options_templates.update(options_section(('ui', "User interface"), {
"show_progressbar": OptionInfo(True, "Show progressbar"),
"show_progress_every_n_steps": OptionInfo(0, "Show image creation progress every N sampling steps. Set 0 to disable.", gr.Slider, {"minimum": 0, "maximum": 32, "step": 1}),
"show_progress_every_n_steps": OptionInfo(0, "Show image creation progress every N sampling steps. Set to 0 to disable. Set to -1 to show after completion of batch.", gr.Slider, {"minimum": -1, "maximum": 32, "step": 1}),
"show_progress_grid": OptionInfo(True, "Show previews of all images generated in a batch as a grid"),
"return_grid": OptionInfo(True, "Show grid in results for web"),
"do_not_show_images": OptionInfo(False, "Do not show any images in results for web"),
@ -378,6 +408,16 @@ class Options:
def __setattr__(self, key, value):
if self.data is not None:
if key in self.data or key in self.data_labels:
assert not cmd_opts.freeze_settings, "changing settings is disabled"
info = opts.data_labels.get(key, None)
comp_args = info.component_args if info else None
if isinstance(comp_args, dict) and comp_args.get('visible', True) is False:
raise RuntimeError(f"not possible to set {key} because it is restricted")
if cmd_opts.hide_ui_dir_config and key in restricted_opts:
raise RuntimeError(f"not possible to set {key} because it is restricted")
self.data[key] = value
return
@ -394,6 +434,8 @@ class Options:
return super(Options, self).__getattribute__(item)
def save(self, filename):
assert not cmd_opts.freeze_settings, "saving settings is disabled"
with open(filename, "w", encoding="utf8") as file:
json.dump(self.data, file, indent=4)
@ -420,11 +462,12 @@ class Options:
if bad_settings > 0:
print(f"The program is likely to not work with bad settings.\nSettings file: {filename}\nEither fix the file, or delete it and restart.", file=sys.stderr)
def onchange(self, key, func):
def onchange(self, key, func, call=True):
item = self.data_labels.get(key)
item.onchange = func
func()
if call:
func()
def dumpjson(self):
d = {k: self.data.get(k, self.data_labels.get(k).default) for k in self.data_labels.keys()}

View file

@ -47,6 +47,8 @@ def txt2img(prompt: str, negative_prompt: str, prompt_style: str, prompt_style2:
if processed is None:
processed = process_images(p)
p.close()
shared.total_tqdm.clear()
generation_info_js = processed.js()

View file

@ -175,9 +175,9 @@ def save_pil_to_file(pil_image, dir=None):
gr.processing_utils.save_pil_to_file = save_pil_to_file
def wrap_gradio_call(func, extra_outputs=None):
def wrap_gradio_call(func, extra_outputs=None, add_stats=False):
def f(*args, extra_outputs_array=extra_outputs, **kwargs):
run_memmon = opts.memmon_poll_rate > 0 and not shared.mem_mon.disabled
run_memmon = opts.memmon_poll_rate > 0 and not shared.mem_mon.disabled and add_stats
if run_memmon:
shared.mem_mon.monitor()
t = time.perf_counter()
@ -204,11 +204,18 @@ def wrap_gradio_call(func, extra_outputs=None):
res = extra_outputs_array + [f"<div class='error'>{plaintext_to_html(type(e).__name__+': '+str(e))}</div>"]
shared.state.skipped = False
shared.state.interrupted = False
shared.state.job_count = 0
if not add_stats:
return tuple(res)
elapsed = time.perf_counter() - t
elapsed_m = int(elapsed // 60)
elapsed_s = elapsed % 60
elapsed_text = f"{elapsed_s:.2f}s"
if (elapsed_m > 0):
if elapsed_m > 0:
elapsed_text = f"{elapsed_m}m "+elapsed_text
if run_memmon:
@ -226,10 +233,6 @@ def wrap_gradio_call(func, extra_outputs=None):
# last item is always HTML
res[-1] += f"<div class='performance'><p class='time'>Time taken: <wbr>{elapsed_text}</p>{vram_html}</div>"
shared.state.skipped = False
shared.state.interrupted = False
shared.state.job_count = 0
return tuple(res)
return f
@ -277,16 +280,8 @@ def check_progress_call(id_part):
image = gr_show(False)
preview_visibility = gr_show(False)
if opts.show_progress_every_n_steps > 0:
if shared.parallel_processing_allowed:
if shared.state.sampling_step - shared.state.current_image_sampling_step >= opts.show_progress_every_n_steps and shared.state.current_latent is not None:
if opts.show_progress_grid:
shared.state.current_image = modules.sd_samplers.samples_to_image_grid(shared.state.current_latent)
else:
shared.state.current_image = modules.sd_samplers.sample_to_image(shared.state.current_latent)
shared.state.current_image_sampling_step = shared.state.sampling_step
if opts.show_progress_every_n_steps != 0:
shared.state.set_current_image()
image = shared.state.current_image
if image is None:
@ -689,6 +684,8 @@ def create_ui(wrap_gradio_gpu_call):
import modules.img2img
import modules.txt2img
reload_javascript()
parameters_copypaste.reset()
with gr.Blocks(analytics_enabled=False) as txt2img_interface:
@ -1083,9 +1080,11 @@ def create_ui(wrap_gradio_gpu_call):
extras_batch_output_dir = gr.Textbox(label="Output directory", **shared.hide_dirs, placeholder="Leave blank to save images to the default path.")
show_extras_results = gr.Checkbox(label='Show result images', value=True)
submit = gr.Button('Generate', elem_id="extras_generate", variant='primary')
with gr.Tabs(elem_id="extras_resize_mode"):
with gr.TabItem('Scale by'):
upscaling_resize = gr.Slider(minimum=1.0, maximum=4.0, step=0.05, label="Resize", value=2)
upscaling_resize = gr.Slider(minimum=1.0, maximum=8.0, step=0.05, label="Resize", value=4)
with gr.TabItem('Scale to'):
with gr.Group():
with gr.Row():
@ -1110,8 +1109,6 @@ def create_ui(wrap_gradio_gpu_call):
with gr.Group():
upscale_before_face_fix = gr.Checkbox(label='Upscale Before Restoring Faces', value=False)
submit = gr.Button('Generate', elem_id="extras_generate", variant='primary')
result_images, html_info_x, html_info = create_output_panel("extras", opts.outdir_extras_samples)
submit.click(
@ -1169,7 +1166,7 @@ def create_ui(wrap_gradio_gpu_call):
outputs=[html, generation_info, html2],
)
with gr.Blocks() as modelmerger_interface:
with gr.Blocks(analytics_enabled=False) as modelmerger_interface:
with gr.Row().style(equal_height=False):
with gr.Column(variant='panel'):
gr.HTML(value="<p>A merger of the two checkpoints will be generated in your <b>checkpoint</b> directory.</p>")
@ -1189,7 +1186,7 @@ def create_ui(wrap_gradio_gpu_call):
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings()
with gr.Blocks() as train_interface:
with gr.Blocks(analytics_enabled=False) as train_interface:
with gr.Row().style(equal_height=False):
gr.HTML(value="<p style='margin-bottom: 0.7em'>See <b><a href=\"https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Textual-Inversion\">wiki</a></b> for detailed explanation.</p>")
@ -1213,8 +1210,8 @@ def create_ui(wrap_gradio_gpu_call):
new_hypernetwork_name = gr.Textbox(label="Name")
new_hypernetwork_sizes = gr.CheckboxGroup(label="Modules", value=["768", "320", "640", "1280"], choices=["768", "320", "640", "1280"])
new_hypernetwork_layer_structure = gr.Textbox("1, 2, 1", label="Enter hypernetwork layer structure", placeholder="1st and last digit must be 1. ex:'1, 2, 1'")
new_hypernetwork_activation_func = gr.Dropdown(value="linear", label="Select activation function of hypernetwork", choices=modules.hypernetworks.ui.keys)
new_hypernetwork_initialization_option = gr.Dropdown(value = "Normal", label="Select Layer weights initialization. relu-like - Kaiming, sigmoid-like - Xavier is recommended", choices=["Normal", "KaimingUniform", "KaimingNormal", "XavierUniform", "XavierNormal"])
new_hypernetwork_activation_func = gr.Dropdown(value="linear", label="Select activation function of hypernetwork. Recommended : Swish / Linear(none)", choices=modules.hypernetworks.ui.keys)
new_hypernetwork_initialization_option = gr.Dropdown(value = "Normal", label="Select Layer weights initialization. Recommended: Kaiming for relu-like, Xavier for sigmoid-like, Normal otherwise", choices=["Normal", "KaimingUniform", "KaimingNormal", "XavierUniform", "XavierNormal"])
new_hypernetwork_add_layer_norm = gr.Checkbox(label="Add layer normalization")
new_hypernetwork_use_dropout = gr.Checkbox(label="Use dropout")
overwrite_old_hypernetwork = gr.Checkbox(value=False, label="Overwrite Old Hypernetwork")
@ -1448,15 +1445,14 @@ def create_ui(wrap_gradio_gpu_call):
if info.refresh is not None:
if is_quicksettings:
res = comp(label=info.label, value=fun, elem_id=elem_id, **(args or {}))
res = comp(label=info.label, value=fun(), elem_id=elem_id, **(args or {}))
create_refresh_button(res, info.refresh, info.component_args, "refresh_" + key)
else:
with gr.Row(variant="compact"):
res = comp(label=info.label, value=fun, elem_id=elem_id, **(args or {}))
res = comp(label=info.label, value=fun(), elem_id=elem_id, **(args or {}))
create_refresh_button(res, info.refresh, info.component_args, "refresh_" + key)
else:
res = comp(label=info.label, value=fun, elem_id=elem_id, **(args or {}))
res = comp(label=info.label, value=fun(), elem_id=elem_id, **(args or {}))
return res
@ -1467,50 +1463,41 @@ def create_ui(wrap_gradio_gpu_call):
opts.reorder()
def run_settings(*args):
changed = 0
assert not shared.cmd_opts.freeze_settings, "changing settings is disabled"
changed = []
for key, value, comp in zip(opts.data_labels.keys(), args, components):
if comp != dummy_component and not opts.same_type(value, opts.data_labels[key].default):
return f"Bad value for setting {key}: {value}; expecting {type(opts.data_labels[key].default).__name__}", opts.dumpjson()
assert comp == dummy_component or opts.same_type(value, opts.data_labels[key].default), f"Bad value for setting {key}: {value}; expecting {type(opts.data_labels[key].default).__name__}"
for key, value, comp in zip(opts.data_labels.keys(), args, components):
if comp == dummy_component:
continue
comp_args = opts.data_labels[key].component_args
if comp_args and isinstance(comp_args, dict) and comp_args.get('visible') is False:
continue
if cmd_opts.hide_ui_dir_config and key in restricted_opts:
continue
oldval = opts.data.get(key, None)
opts.data[key] = value
try:
setattr(opts, key, value)
except RuntimeError:
continue
if oldval != value:
if opts.data_labels[key].onchange is not None:
opts.data_labels[key].onchange()
changed += 1
opts.save(shared.config_filename)
return f'{changed} settings changed.', opts.dumpjson()
changed.append(key)
try:
opts.save(shared.config_filename)
except RuntimeError:
return opts.dumpjson(), f'{len(changed)} settings changed without save: {", ".join(changed)}.'
return opts.dumpjson(), f'{len(changed)} settings changed: {", ".join(changed)}.'
def run_settings_single(value, key):
assert not shared.cmd_opts.freeze_settings, "changing settings is disabled"
if not opts.same_type(value, opts.data_labels[key].default):
return gr.update(visible=True), opts.dumpjson()
oldval = opts.data.get(key, None)
if cmd_opts.hide_ui_dir_config and key in restricted_opts:
try:
setattr(opts, key, value)
except Exception:
return gr.update(value=oldval), opts.dumpjson()
opts.data[key] = value
if oldval != value:
if opts.data_labels[key].onchange is not None:
opts.data_labels[key].onchange()
@ -1595,8 +1582,7 @@ def create_ui(wrap_gradio_gpu_call):
reload_script_bodies.click(
fn=reload_scripts,
inputs=[],
outputs=[],
_js='function(){}'
outputs=[]
)
def request_restart():
@ -1604,11 +1590,10 @@ def create_ui(wrap_gradio_gpu_call):
shared.state.need_restart = True
restart_gradio.click(
fn=request_restart,
_js='restart_reload',
inputs=[],
outputs=[],
_js='function(){restart_reload()}'
)
if column is not None:
@ -1664,9 +1649,9 @@ def create_ui(wrap_gradio_gpu_call):
text_settings = gr.Textbox(elem_id="settings_json", value=lambda: opts.dumpjson(), visible=False)
settings_submit.click(
fn=run_settings,
fn=wrap_gradio_call(run_settings, extra_outputs=[gr.update()]),
inputs=components,
outputs=[result, text_settings],
outputs=[text_settings, result],
)
for i, k, item in quicksettings_list:
@ -1678,6 +1663,17 @@ def create_ui(wrap_gradio_gpu_call):
outputs=[component, text_settings],
)
component_keys = [k for k in opts.data_labels.keys() if k in component_dict]
def get_settings_values():
return [getattr(opts, key) for key in component_keys]
demo.load(
fn=get_settings_values,
inputs=[],
outputs=[component_dict[k] for k in component_keys],
)
def modelmerger(*args):
try:
results = modules.extras.run_modelmerger(*args)
@ -1807,4 +1803,3 @@ def load_javascript(raw_response):
reload_javascript = partial(load_javascript, gradio.routes.templates.TemplateResponse)
reload_javascript()

View file

@ -86,7 +86,7 @@ def extension_table():
code += f"""
<tr>
<td><label><input class="gr-check-radio gr-checkbox" name="enable_{html.escape(ext.name)}" type="checkbox" {'checked="checked"' if ext.enabled else ''}>{html.escape(ext.name)}</label></td>
<td><a href="{html.escape(ext.remote or '')}">{html.escape(ext.remote or '')}</a></td>
<td><a href="{html.escape(ext.remote or '')}" target="_blank">{html.escape(ext.remote or '')}</a></td>
<td{' class="extension_status"' if ext.remote is not None else ''}>{ext_status}</td>
</tr>
"""
@ -140,13 +140,15 @@ def install_extension_from_url(dirname, url):
shutil.rmtree(tmpdir, True)
def install_extension_from_index(url):
def install_extension_from_index(url, hide_tags):
ext_table, message = install_extension_from_url(None, url)
return refresh_available_extensions_from_data(), ext_table, message
code, _ = refresh_available_extensions_from_data(hide_tags)
return code, ext_table, message
def refresh_available_extensions(url):
def refresh_available_extensions(url, hide_tags):
global available_extensions
import urllib.request
@ -155,13 +157,25 @@ def refresh_available_extensions(url):
available_extensions = json.loads(text)
return url, refresh_available_extensions_from_data(), ''
code, tags = refresh_available_extensions_from_data(hide_tags)
return url, code, gr.CheckboxGroup.update(choices=tags), ''
def refresh_available_extensions_from_data():
def refresh_available_extensions_for_tags(hide_tags):
code, _ = refresh_available_extensions_from_data(hide_tags)
return code, ''
def refresh_available_extensions_from_data(hide_tags):
extlist = available_extensions["extensions"]
installed_extension_urls = {normalize_git_url(extension.remote): extension.name for extension in extensions.extensions}
tags = available_extensions.get("tags", {})
tags_to_hide = set(hide_tags)
hidden = 0
code = f"""<!-- {time.time()} -->
<table id="available_extensions">
<thead>
@ -178,17 +192,24 @@ def refresh_available_extensions_from_data():
name = ext.get("name", "noname")
url = ext.get("url", None)
description = ext.get("description", "")
extension_tags = ext.get("tags", [])
if url is None:
continue
if len([x for x in extension_tags if x in tags_to_hide]) > 0:
hidden += 1
continue
existing = installed_extension_urls.get(normalize_git_url(url), None)
install_code = f"""<input onclick="install_extension_from_index(this, '{html.escape(url)}')" type="button" value="{"Install" if not existing else "Installed"}" {"disabled=disabled" if existing else ""} class="gr-button gr-button-lg gr-button-secondary">"""
tags_text = ", ".join([f"<span class='extension-tag' title='{tags.get(x, '')}'>{x}</span>" for x in extension_tags])
code += f"""
<tr>
<td><a href="{html.escape(url)}">{html.escape(name)}</a></td>
<td><a href="{html.escape(url)}" target="_blank">{html.escape(name)}</a><br />{tags_text}</td>
<td>{html.escape(description)}</td>
<td>{install_code}</td>
</tr>
@ -199,7 +220,10 @@ def refresh_available_extensions_from_data():
</table>
"""
return code
if hidden > 0:
code += f"<p>Extension hidden: {hidden}</p>"
return code, list(tags)
def create_ui():
@ -238,21 +262,30 @@ def create_ui():
extension_to_install = gr.Text(elem_id="extension_to_install", visible=False)
install_extension_button = gr.Button(elem_id="install_extension_button", visible=False)
with gr.Row():
hide_tags = gr.CheckboxGroup(value=["ads", "localization"], label="Hide extensions with tags", choices=["script", "ads", "localization"])
install_result = gr.HTML()
available_extensions_table = gr.HTML()
refresh_available_extensions_button.click(
fn=modules.ui.wrap_gradio_call(refresh_available_extensions, extra_outputs=[gr.update(), gr.update()]),
inputs=[available_extensions_index],
outputs=[available_extensions_index, available_extensions_table, install_result],
fn=modules.ui.wrap_gradio_call(refresh_available_extensions, extra_outputs=[gr.update(), gr.update(), gr.update()]),
inputs=[available_extensions_index, hide_tags],
outputs=[available_extensions_index, available_extensions_table, hide_tags, install_result],
)
install_extension_button.click(
fn=modules.ui.wrap_gradio_call(install_extension_from_index, extra_outputs=[gr.update(), gr.update()]),
inputs=[extension_to_install],
inputs=[extension_to_install, hide_tags],
outputs=[available_extensions_table, extensions_table, install_result],
)
hide_tags.change(
fn=modules.ui.wrap_gradio_call(refresh_available_extensions_for_tags, extra_outputs=[gr.update()]),
inputs=[hide_tags],
outputs=[available_extensions_table, install_result]
)
with gr.TabItem("Install from URL"):
install_url = gr.Text(label="URL for extension's git repository")
install_dirname = gr.Text(label="Local directory name", placeholder="Leave empty for auto")

View file

@ -10,6 +10,7 @@ import modules.shared
from modules import modelloader, shared
LANCZOS = (Image.Resampling.LANCZOS if hasattr(Image, 'Resampling') else Image.LANCZOS)
NEAREST = (Image.Resampling.NEAREST if hasattr(Image, 'Resampling') else Image.NEAREST)
from modules.paths import models_path
@ -56,10 +57,18 @@ class Upscaler:
self.scale = scale
dest_w = img.width * scale
dest_h = img.height * scale
for i in range(3):
shape = (img.width, img.height)
img = self.do_upscale(img, selected_model)
if shape == (img.width, img.height):
break
if img.width >= dest_w and img.height >= dest_h:
break
img = self.do_upscale(img, selected_model)
if img.width != dest_w or img.height != dest_h:
img = img.resize((int(dest_w), int(dest_h)), resample=LANCZOS)
@ -120,3 +129,17 @@ class UpscalerLanczos(Upscaler):
self.name = "Lanczos"
self.scalers = [UpscalerData("Lanczos", None, self)]
class UpscalerNearest(Upscaler):
scalers = []
def do_upscale(self, img, selected_model=None):
return img.resize((int(img.width * self.scale), int(img.height * self.scale)), resample=NEAREST)
def load_model(self, _):
pass
def __init__(self, dirname=None):
super().__init__(False)
self.name = "Nearest"
self.scalers = [UpscalerData("Nearest", None, self)]

View file

@ -4,7 +4,7 @@ fairscale==0.4.4
fonts
font-roboto
gfpgan
gradio==3.5
gradio==3.9
invisible-watermark
numpy
omegaconf

View file

@ -2,7 +2,7 @@ transformers==4.19.2
diffusers==0.3.0
basicsr==1.4.2
gfpgan==1.3.8
gradio==3.5
gradio==3.9
numpy==1.23.3
Pillow==9.2.0
realesrgan==0.3.0

View file

@ -14,7 +14,7 @@ class Script(scripts.Script):
return cmd_opts.allow_code
def ui(self, is_img2img):
code = gr.Textbox(label="Python code", visible=False, lines=1)
code = gr.Textbox(label="Python code", lines=1)
return [code]

View file

@ -166,8 +166,7 @@ class Script(scripts.Script):
if override_strength:
p.denoising_strength = 1.0
def sample_extra(conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength):
def sample_extra(conditioning, unconditional_conditioning, seeds, subseeds, subseed_strength, prompts):
lat = (p.init_latent.cpu().numpy() * 10).astype(int)
same_params = self.cache is not None and self.cache.cfg_scale == cfg and self.cache.steps == st \

View file

@ -132,7 +132,7 @@ class Script(scripts.Script):
info = gr.HTML("<p style=\"margin-bottom:0.75em\">Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8</p>")
pixels = gr.Slider(label="Pixels to expand", minimum=8, maximum=256, step=8, value=128)
mask_blur = gr.Slider(label='Mask blur', minimum=0, maximum=64, step=1, value=8, visible=False)
mask_blur = gr.Slider(label='Mask blur', minimum=0, maximum=64, step=1, value=8)
direction = gr.CheckboxGroup(label="Outpainting direction", choices=['left', 'right', 'up', 'down'], value=['left', 'right', 'up', 'down'])
noise_q = gr.Slider(label="Fall-off exponent (lower=higher detail)", minimum=0.0, maximum=4.0, step=0.01, value=1.0)
color_variation = gr.Slider(label="Color variation", minimum=0.0, maximum=1.0, step=0.01, value=0.05)

View file

@ -22,8 +22,8 @@ class Script(scripts.Script):
return None
pixels = gr.Slider(label="Pixels to expand", minimum=8, maximum=256, step=8, value=128)
mask_blur = gr.Slider(label='Mask blur', minimum=0, maximum=64, step=1, value=4, visible=False)
inpainting_fill = gr.Radio(label='Masked content', choices=['fill', 'original', 'latent noise', 'latent nothing'], value='fill', type="index", visible=False)
mask_blur = gr.Slider(label='Mask blur', minimum=0, maximum=64, step=1, value=4)
inpainting_fill = gr.Radio(label='Masked content', choices=['fill', 'original', 'latent noise', 'latent nothing'], value='fill', type="index")
direction = gr.CheckboxGroup(label="Outpainting direction", choices=['left', 'right', 'up', 'down'], value=['left', 'right', 'up', 'down'])
return [pixels, mask_blur, inpainting_fill, direction]

View file

@ -83,19 +83,21 @@ def cmdargs(line):
def load_prompt_file(file):
if (file is None):
if file is None:
lines = []
else:
lines = [x.strip() for x in file.decode('utf8', errors='ignore').split("\n")]
return None, "\n".join(lines), gr.update(lines=7)
class Script(scripts.Script):
def title(self):
return "Prompts from file or textbox"
def ui(self, is_img2img):
checkbox_iterate = gr.Checkbox(label="Iterate seed every line", value=False)
checkbox_iterate_batch = gr.Checkbox(label="Use same random seed for all lines", value=False)
prompt_txt = gr.Textbox(label="List of prompt inputs", lines=1)
file = gr.File(label="Upload prompt inputs", type='bytes')
@ -106,9 +108,9 @@ class Script(scripts.Script):
# We don't shrink back to 1, because that causes the control to ignore [enter], and it may
# be unclear to the user that shift-enter is needed.
prompt_txt.change(lambda tb: gr.update(lines=7) if ("\n" in tb) else gr.update(lines=2), inputs=[prompt_txt], outputs=[prompt_txt])
return [checkbox_iterate, file, prompt_txt]
return [checkbox_iterate, checkbox_iterate_batch, prompt_txt]
def run(self, p, checkbox_iterate, file, prompt_txt: str):
def run(self, p, checkbox_iterate, checkbox_iterate_batch, prompt_txt: str):
lines = [x.strip() for x in prompt_txt.splitlines()]
lines = [x for x in lines if len(x) > 0]
@ -137,7 +139,7 @@ class Script(scripts.Script):
jobs.append(args)
print(f"Will process {len(lines)} lines in {job_count} jobs.")
if (checkbox_iterate and p.seed == -1):
if (checkbox_iterate or checkbox_iterate_batch) and p.seed == -1:
p.seed = int(random.randrange(4294967294))
state.job_count = job_count
@ -153,8 +155,7 @@ class Script(scripts.Script):
proc = process_images(copy_p)
images += proc.images
if (checkbox_iterate):
if checkbox_iterate:
p.seed = p.seed + (p.batch_size * p.n_iter)
return Processed(p, images, p.seed, "")
return Processed(p, images, p.seed, "")

View file

@ -18,8 +18,8 @@ class Script(scripts.Script):
def ui(self, is_img2img):
info = gr.HTML("<p style=\"margin-bottom:0.75em\">Will upscale the image to twice the dimensions; use width and height sliders to set tile size</p>")
overlap = gr.Slider(minimum=0, maximum=256, step=16, label='Tile overlap', value=64, visible=False)
upscaler_index = gr.Radio(label='Upscaler', choices=[x.name for x in shared.sd_upscalers], value=shared.sd_upscalers[0].name, type="index", visible=False)
overlap = gr.Slider(minimum=0, maximum=256, step=16, label='Tile overlap', value=64)
upscaler_index = gr.Radio(label='Upscaler', choices=[x.name for x in shared.sd_upscalers], value=shared.sd_upscalers[0].name, type="index")
return [info, overlap, upscaler_index]

View file

@ -263,12 +263,12 @@ class Script(scripts.Script):
current_axis_options = [x for x in axis_options if type(x) == AxisOption or type(x) == AxisOptionImg2Img and is_img2img]
with gr.Row():
x_type = gr.Dropdown(label="X type", choices=[x.label for x in current_axis_options], value=current_axis_options[1].label, visible=False, type="index", elem_id="x_type")
x_values = gr.Textbox(label="X values", visible=False, lines=1)
x_type = gr.Dropdown(label="X type", choices=[x.label for x in current_axis_options], value=current_axis_options[1].label, type="index", elem_id="x_type")
x_values = gr.Textbox(label="X values", lines=1)
with gr.Row():
y_type = gr.Dropdown(label="Y type", choices=[x.label for x in current_axis_options], value=current_axis_options[0].label, visible=False, type="index", elem_id="y_type")
y_values = gr.Textbox(label="Y values", visible=False, lines=1)
y_type = gr.Dropdown(label="Y type", choices=[x.label for x in current_axis_options], value=current_axis_options[0].label, type="index", elem_id="y_type")
y_values = gr.Textbox(label="Y values", lines=1)
draw_legend = gr.Checkbox(label='Draw legend', value=True)
include_lone_images = gr.Checkbox(label='Include Separate Images', value=False)

View file

@ -260,6 +260,16 @@ input[type="range"]{
#txt2img_negative_prompt, #img2img_negative_prompt{
}
/* gradio 3.8 adds opacity to progressbar which makes it blink; disable it here */
.transition.opacity-20 {
opacity: 1 !important;
}
/* more gradio's garbage cleanup */
.min-h-\[4rem\] {
min-height: unset !important;
}
#txt2img_progressbar, #img2img_progressbar, #ti_progressbar{
position: absolute;
z-index: 1000;
@ -491,7 +501,7 @@ input[type="range"]{
padding: 0;
}
#refresh_sd_model_checkpoint, #refresh_sd_hypernetwork, #refresh_train_hypernetwork_name, #refresh_train_embedding_name, #refresh_localization{
#refresh_sd_model_checkpoint, #refresh_sd_vae, #refresh_sd_hypernetwork, #refresh_train_hypernetwork_name, #refresh_train_embedding_name, #refresh_localization{
max-width: 2.5em;
min-width: 2.5em;
height: 2.4em;
@ -553,6 +563,11 @@ img2maskimg, #img2maskimg > .h-60, #img2maskimg > .h-60 > div, #img2maskimg > .h
opacity: 0.5;
}
.extension-tag{
font-weight: bold;
font-size: 95%;
}
/* The following handles localization for right-to-left (RTL) languages like Arabic.
The rtl media type will only be activated by the logic in javascript/localization.js.
If you change anything above, you need to make sure it is RTL compliant by just running

63
test/utils_test.py Normal file
View file

@ -0,0 +1,63 @@
import unittest
import requests
class UtilsTests(unittest.TestCase):
def setUp(self):
self.url_options = "http://localhost:7860/sdapi/v1/options"
self.url_cmd_flags = "http://localhost:7860/sdapi/v1/cmd-flags"
self.url_samplers = "http://localhost:7860/sdapi/v1/samplers"
self.url_upscalers = "http://localhost:7860/sdapi/v1/upscalers"
self.url_sd_models = "http://localhost:7860/sdapi/v1/sd-models"
self.url_hypernetworks = "http://localhost:7860/sdapi/v1/hypernetworks"
self.url_face_restorers = "http://localhost:7860/sdapi/v1/face-restorers"
self.url_realesrgan_models = "http://localhost:7860/sdapi/v1/realesrgan-models"
self.url_prompt_styles = "http://localhost:7860/sdapi/v1/prompt-styles"
self.url_artist_categories = "http://localhost:7860/sdapi/v1/artist-categories"
self.url_artists = "http://localhost:7860/sdapi/v1/artists"
def test_options_get(self):
self.assertEqual(requests.get(self.url_options).status_code, 200)
def test_options_write(self):
response = requests.get(self.url_options)
self.assertEqual(response.status_code, 200)
pre_value = response.json()["send_seed"]
self.assertEqual(requests.post(self.url_options, json={"send_seed":not pre_value}).status_code, 200)
response = requests.get(self.url_options)
self.assertEqual(response.status_code, 200)
self.assertEqual(response.json()["send_seed"], not pre_value)
requests.post(self.url_options, json={"send_seed": pre_value})
def test_cmd_flags(self):
self.assertEqual(requests.get(self.url_cmd_flags).status_code, 200)
def test_samplers(self):
self.assertEqual(requests.get(self.url_samplers).status_code, 200)
def test_upscalers(self):
self.assertEqual(requests.get(self.url_upscalers).status_code, 200)
def test_sd_models(self):
self.assertEqual(requests.get(self.url_sd_models).status_code, 200)
def test_hypernetworks(self):
self.assertEqual(requests.get(self.url_hypernetworks).status_code, 200)
def test_face_restorers(self):
self.assertEqual(requests.get(self.url_face_restorers).status_code, 200)
def test_realesrgan_models(self):
self.assertEqual(requests.get(self.url_realesrgan_models).status_code, 200)
def test_prompt_styles(self):
self.assertEqual(requests.get(self.url_prompt_styles).status_code, 200)
def test_artist_categories(self):
self.assertEqual(requests.get(self.url_artist_categories).status_code, 200)
def test_artists(self):
self.assertEqual(requests.get(self.url_artists).status_code, 200)

View file

@ -5,11 +5,12 @@ import importlib
import signal
import threading
from fastapi import FastAPI
from fastapi.middleware.cors import CORSMiddleware
from fastapi.middleware.gzip import GZipMiddleware
from modules.paths import script_path
from modules import devices, sd_samplers, upscaler, extensions
from modules import devices, sd_samplers, upscaler, extensions, localization
import modules.codeformer_model as codeformer
import modules.extras
import modules.face_restoration
@ -21,19 +22,18 @@ import modules.paths
import modules.scripts
import modules.sd_hijack
import modules.sd_models
import modules.sd_vae
import modules.shared as shared
import modules.txt2img
import modules.script_callbacks
import modules.ui
from modules import devices
from modules import modelloader
from modules.paths import script_path
from modules.shared import cmd_opts
import modules.hypernetworks.hypernetwork
queue_lock = threading.Lock()
server_name = "0.0.0.0" if cmd_opts.listen else cmd_opts.server_name
def wrap_queued_call(func):
def f(*args, **kwargs):
@ -57,11 +57,12 @@ def wrap_gradio_gpu_call(func, extra_outputs=None):
return res
return modules.ui.wrap_gradio_call(f, extra_outputs=extra_outputs)
return modules.ui.wrap_gradio_call(f, extra_outputs=extra_outputs, add_stats=True)
def initialize():
extensions.list_extensions()
localization.list_localizations(cmd_opts.localizations_dir)
if cmd_opts.ui_debug_mode:
shared.sd_upscalers = upscaler.UpscalerLanczos().scalers
@ -77,11 +78,26 @@ def initialize():
modules.scripts.load_scripts()
modules.sd_vae.refresh_vae_list()
modules.sd_models.load_model()
shared.opts.onchange("sd_model_checkpoint", wrap_queued_call(lambda: modules.sd_models.reload_model_weights()))
shared.opts.onchange("sd_vae", wrap_queued_call(lambda: modules.sd_vae.reload_vae_weights()), call=False)
shared.opts.onchange("sd_hypernetwork", wrap_queued_call(lambda: modules.hypernetworks.hypernetwork.load_hypernetwork(shared.opts.sd_hypernetwork)))
shared.opts.onchange("sd_hypernetwork_strength", modules.hypernetworks.hypernetwork.apply_strength)
if cmd_opts.tls_keyfile is not None and cmd_opts.tls_keyfile is not None:
try:
if not os.path.exists(cmd_opts.tls_keyfile):
print("Invalid path to TLS keyfile given")
if not os.path.exists(cmd_opts.tls_certfile):
print(f"Invalid path to TLS certfile: '{cmd_opts.tls_certfile}'")
except TypeError:
cmd_opts.tls_keyfile = cmd_opts.tls_certfile = None
print("TLS setup invalid, running webui without TLS")
else:
print("Running with TLS")
# make the program just exit at ctrl+c without waiting for anything
def sigint_handler(sig, frame):
print(f'Interrupted with signal {sig} in {frame}')
@ -90,6 +106,11 @@ def initialize():
signal.signal(signal.SIGINT, sigint_handler)
def setup_cors(app):
if cmd_opts.cors_allow_origins:
app.add_middleware(CORSMiddleware, allow_origins=cmd_opts.cors_allow_origins.split(','), allow_methods=['*'])
def create_api(app):
from modules.api.api import Api
api = Api(app, queue_lock)
@ -111,9 +132,12 @@ def api_only():
initialize()
app = FastAPI()
setup_cors(app)
app.add_middleware(GZipMiddleware, minimum_size=1000)
api = create_api(app)
modules.script_callbacks.app_started_callback(None, app)
api.launch(server_name="0.0.0.0" if cmd_opts.listen else "127.0.0.1", port=cmd_opts.port if cmd_opts.port else 7861)
@ -126,8 +150,10 @@ def webui():
app, local_url, share_url = demo.launch(
share=cmd_opts.share,
server_name="0.0.0.0" if cmd_opts.listen else None,
server_name=server_name,
server_port=cmd_opts.port,
ssl_keyfile=cmd_opts.tls_keyfile,
ssl_certfile=cmd_opts.tls_certfile,
debug=cmd_opts.gradio_debug,
auth=[tuple(cred.split(':')) for cred in cmd_opts.gradio_auth.strip('"').split(',')] if cmd_opts.gradio_auth else None,
inbrowser=cmd_opts.autolaunch,
@ -136,6 +162,14 @@ def webui():
# after initial launch, disable --autolaunch for subsequent restarts
cmd_opts.autolaunch = False
# gradio uses a very open CORS policy via app.user_middleware, which makes it possible for
# an attacker to trick the user into opening a malicious HTML page, which makes a request to the
# running web ui and do whatever the attcker wants, including installing an extension and
# runnnig its code. We disable this here. Suggested by RyotaK.
app.user_middleware = [x for x in app.user_middleware if x.cls.__name__ != 'CORSMiddleware']
setup_cors(app)
app.add_middleware(GZipMiddleware, minimum_size=1000)
if launch_api:
@ -149,6 +183,9 @@ def webui():
print('Reloading extensions')
extensions.list_extensions()
localization.list_localizations(cmd_opts.localizations_dir)
print('Reloading custom scripts')
modules.scripts.reload_scripts()
print('Reloading modules: modules.ui')