AI Image & Video Generation Glossary

Over 100 terms from the world of AI image and video generation with audio, explained in plain language. Whether you are just getting started with AI image generation or fine-tuning advanced workflows, this glossary covers the vocabulary you need.

What do AI image generation terms mean? This glossary explains over 100 key AI image and video generation with audio terms in plain language, including diffusion models, photorealistic, stylized, LoRA, CFG scale, ControlNet, latent space, and more. Each definition is written for both beginners and advanced users working with tools like ZSky AI, Stable Diffusion, and Midjourney.

ABCDEFGHIJKLMNOPQRSTUVWXZ
A

Aspect Ratio

The proportional relationship between an image's width and height, expressed as two numbers separated by a colon (e.g., 16:9, 1:1, 9:16). Choosing the right aspect ratio is essential for matching output to its intended use, whether social media posts, widescreen video, or print. Learn about prompting for specific ratios.

Attention Mechanism

A neural network component that allows the model to focus on the most relevant parts of input data when generating output. In image generation, attention mechanisms help the model understand which words in a prompt correspond to which visual regions, enabling precise compositional control.

Autoencoder

A neural network that learns to compress data into a compact latent representation and then reconstruct it. Autoencoders are a foundational component in diffusion model pipelines, where they encode images to latent space for efficient processing and decode them back to full resolution.

AI Art Generator

Software that uses artificial intelligence models to create visual artwork from text descriptions, reference images, or other inputs. Modern AI art generators like ZSky AI use diffusion models such as advanced AI to produce high-quality results. How AI image generation works.

AI Upscaling

The process of using neural networks to increase an image's resolution while adding realistic detail that was not present in the original. Unlike simple interpolation, AI upscaling models like ESRGAN predict and synthesize new pixels based on learned patterns from training data. Try ZSky AI's upscaler.

Alpha Channel

A fourth channel in image data (beyond RGB) that stores transparency information. Alpha channels are important for AI-generated images used in compositing, allowing subjects to be placed on any background without visible edges.

Anime Style

A visual aesthetic inspired by Japanese animation, characterized by bold outlines, flat color areas, and expressive characters. Many AI models include anime-specific training data or can be fine-tuned with LoRA adapters for anime output. Generate anime art with ZSky AI.

B

Batch Size

The number of images generated simultaneously in a single generation run. Larger batch sizes allow you to explore more variations of a prompt at once but require more GPU memory. ZSky AI processes batches efficiently on dedicated RTX 5090 hardware.

Blackwell Architecture

NVIDIA's GPU architecture powering the RTX 5090 series. It introduces significant improvements in AI inference throughput, memory bandwidth, and power efficiency compared to previous generations, making it ideal for running large diffusion models at full precision.

Blending

A technique for combining two or more images or prompts together to create hybrid results. Blending can merge visual styles, subjects, or compositions, producing creative outputs that draw from multiple source concepts simultaneously.

BLIP

Bootstrapping Language-Image Pre-training, a model architecture for understanding the relationship between images and text. BLIP and its successors are used for image captioning, visual question answering, and generating text descriptions from images.

C

Canny Edge Detection

An image processing algorithm that detects structural edges and outlines in images. In AI generation workflows, canny edge maps serve as ControlNet inputs to guide the model in replicating the structure of a reference image while generating new visual content.

CFG Scale (Classifier-Free Guidance)

A parameter that controls how strictly the AI follows your text prompt during generation. Higher CFG values (7-15) produce images more closely matching the prompt but may appear over-saturated, while lower values (1-5) give the model more creative latitude. Prompt writing guide.

Checkpoint

A saved snapshot of a fully trained AI model's weights at a specific point in training. Checkpoints are the base model files you load to generate images. Different checkpoints produce different visual styles and capabilities, from photorealism to anime to illustration.

CLIP (Contrastive Language-Image Pre-training)

A neural network trained by OpenAI that understands the relationship between text descriptions and images. CLIP serves as the text encoder in many diffusion models, translating your written prompt into numerical representations the image generator can understand.

Color Grading

The process of adjusting the overall color palette and tonal balance of an image or video. AI models can apply color grading through prompt descriptions like "warm tones," "cinematic color grading," or "teal and orange" to achieve specific visual moods.

Compositing

The process of combining multiple visual elements into a single coherent image. In AI workflows, compositing often involves generating separate elements (subjects, backgrounds, effects) and layering them together using tools like inpainting and outpainting.

ControlNet

A neural network architecture that adds precise spatial conditioning to diffusion models. ControlNet accepts structural inputs like depth maps, edge detection, human poses, and segmentation maps to guide image generation while preserving compositional control. It enables workflows like pose-matching and architectural rendering. Learn more about AI generation techniques.

Cross-Attention

An attention mechanism where the model relates information from two different sources, typically the text prompt and the image being generated. Cross-attention layers are how diffusion models connect words in your prompt to specific visual regions in the output.

D

DALL-E

A series of AI image generation models created by OpenAI. DALL-E 3 integrates with ChatGPT and produces high-quality images from text prompts. While powerful, it requires an OpenAI subscription and has strict content policies. Compare with ZSky AI.

Denoising

The core process in diffusion models where random noise is progressively removed from an image over multiple steps. Each denoising step refines the image slightly, guided by the text prompt and model weights, until a clean final image emerges from what started as pure static.

Denoising Strength

A parameter in img2img workflows that controls how much the AI changes the input image. A value of 0.0 returns the original image unchanged, while 1.0 completely replaces it. Values between 0.3-0.7 are typical for modifying images while retaining their core structure.

Depth Map

A grayscale image where pixel brightness represents distance from the camera, with white being nearest and black being farthest. Depth maps serve as ControlNet inputs to guide AI models in creating images with specific spatial arrangements and perspective.

Diffusion Model

A class of generative AI model that creates images by learning to reverse a noise-adding process. During training, the model learns how noise is added to images; during generation, it runs this process in reverse, starting from random noise and progressively denoising it into a coherent image guided by a text prompt. photorealistic, stylized, and Stable Diffusion are all diffusion models. How diffusion models work.

DPI (Dots Per Inch)

A measurement of print resolution indicating how many ink dots fit within one linear inch. For print-quality AI-generated images, 300 DPI is the standard target. Higher generation resolutions and AI upscaling help achieve print-ready DPI. AI images for print.

DreamBooth

A fine-tuning technique that teaches an AI model to generate images of a specific subject (a person, pet, object, or style) from just a few reference photos. DreamBooth creates a persistent concept in the model's knowledge that can be combined with any prompt.

E

Embedding

A numerical vector representation of a concept (text, image, or other data) in a high-dimensional space. In AI image generation, text embeddings capture the meaning of your prompt, while image embeddings represent visual concepts. Similar concepts have similar embeddings, enabling the model to understand relationships.

ESRGAN (Enhanced Super-Resolution GAN)

A neural network architecture specifically designed for image upscaling that produces sharp, detailed results even at 4x magnification. ESRGAN and its variants are the backbone of most AI upscaling tools, restoring fine textures and details that simpler methods cannot. Try AI upscaling.

Euler Sampler

A numerical method used in diffusion model sampling that calculates each denoising step by following the gradient of the learned noise prediction. Euler and Euler Ancestral are among the fastest samplers, often producing good results in fewer steps than more complex alternatives.

EMA (Exponential Moving Average)

A training technique that maintains a smoothed average of model weights over time, producing more stable and higher-quality model checkpoints. EMA weights often generate better images than raw training weights.

F

Face Restoration

An AI post-processing technique that detects faces in generated images and enhances them for improved realism, correcting common artifacts like asymmetrical features, blurred eyes, or distorted expressions. Popular face restoration models include GFPGAN and CodeFormer.

Fine-Tuning

The process of further training a pre-trained AI model on a specific dataset to adapt it to a particular style, subject, or domain. Fine-tuning methods range from full model retraining to lightweight approaches like LoRA and DreamBooth that modify only a fraction of parameters.

Flow Matching

A generative modeling technique that learns continuous transformation paths between noise and data distributions. photorealistic uses flow matching rather than traditional diffusion, enabling more efficient and higher-quality image generation with fewer sampling steps.

photorealistic

A state-of-the-art open-weight image generation model developed by Black Forest Labs. photorealistic is known for exceptional photorealism, accurate text rendering within images, and strong prompt adherence. It uses a transformer-based architecture with flow matching rather than traditional diffusion. Generate with photorealistic on ZSky AI. What is photorealistic AI?

next-gen AI

The second-generation AI engine offering improved quality, faster inference, and expanded capabilities over the original. next-gen AI builds on the same transformer architecture with refined training and additional features. Try next-gen AI on ZSky AI.

FP16 / FP32 (Floating Point Precision)

Numerical precision formats used during AI model inference. FP32 (32-bit) offers maximum precision but uses more memory; FP16 (16-bit) halves memory requirements with minimal quality loss. Most image generation runs at FP16, while some operations benefit from FP32 for accuracy.

Frame Interpolation

A video AI technique that generates intermediate frames between existing ones, creating smoother motion or slow-motion effects. Modern AI frame interpolation models predict complex motion including occlusion and deformation, producing results superior to traditional optical flow methods. AI video generation with audio.

G

GAN (Generative Adversarial Network)

A model architecture where two neural networks compete: a generator that creates images and a discriminator that evaluates their realism. While largely superseded by diffusion models for image generation, GANs remain important for tasks like super-resolution (ESRGAN) and face restoration (GFPGAN). AI generation models explained.

GDDR7 Memory

The latest generation of graphics memory used in GPUs like the RTX 5090. GDDR7 offers significantly higher bandwidth than GDDR6X, enabling faster processing of large AI models and higher-resolution image generation without bottlenecks.

Guidance Scale

See CFG Scale. The parameter that controls prompt adherence during image generation, balancing between creative freedom and strict prompt following.

GPU (Graphics Processing Unit)

Specialized hardware originally designed for rendering graphics, now essential for AI model inference and training. AI image generation relies heavily on GPU parallel processing capabilities. ZSky AI runs on dedicated NVIDIA RTX 5090 GPUs. Dedicated GPU generation.

H

HDR (High Dynamic Range)

An imaging technique that captures or simulates a wider range of brightness levels than standard images. In AI generation, HDR prompts produce images with enhanced contrast, vibrant highlights, and deep shadows that mimic the look of HDR photography.

Hypernetwork

A small auxiliary neural network that modifies the behavior of a larger model by adjusting its cross-attention layers. Hypernetworks were an early fine-tuning approach for Stable Diffusion that allowed style and subject customization without retraining the full model. They have been largely superseded by LoRA.

Hallucination

When an AI model generates content that is visually plausible but factually incorrect or physically impossible, such as hands with extra fingers, text with misspelled words, or objects with impossible geometry. Modern models like photorealistic have significantly reduced hallucination compared to earlier generations.

Hires Fix

A two-pass generation technique where the image is first generated at a lower resolution, then upscaled and refined with additional denoising at the target resolution. This avoids compositional artifacts that can occur when generating directly at high resolutions.

I

Image-to-Image (img2img)

A generation mode where an existing image serves as the starting point, combined with a text prompt to guide modifications. The AI adds noise to the input image and then denoises it according to the prompt, producing a result that blends the original structure with new creative direction. Image to video on ZSky AI.

Inpainting

A technique for selectively regenerating specific areas of an image while leaving the rest unchanged. You mask the area you want to modify, provide a prompt, and the AI fills in only the masked region, seamlessly blending with the surrounding content. Useful for fixing artifacts, changing objects, or extending compositions.

Inference

The process of running a trained AI model to generate output, as opposed to training the model. When you generate an image on ZSky AI, the GPU is performing inference, running your prompt through the model's weights to produce a result.

IP-Adapter (Image Prompt Adapter)

A technique that allows using an image as part of your prompt input, enabling the AI to replicate visual elements like style, composition, or subject appearance from a reference image. IP-Adapter bridges the gap between text prompting and visual reference-based generation.

J

JPEG Artifact

Visual distortions caused by lossy JPEG compression, appearing as blocky patterns and color banding especially around edges and in gradient areas. AI upscalers and enhancement models can remove JPEG artifacts while restoring detail. Generating at PNG format avoids these artifacts entirely.

K

Keyframe

In AI video generation with audio, a keyframe is a user-specified or AI-determined frame that defines a critical moment in the video sequence. The AI generates intermediate frames to smoothly transition between keyframes. Well-chosen keyframes give you compositional control over the video's narrative arc. AI video generation with audio explained.

Kling

An AI video generation with audio model developed by Kuaishou Technology, known for producing high-quality video clips with realistic motion. Kling competes in the AI video space alongside Sora, Runway, and Pika. ZSky AI as Kling alternative.

KSampler

A family of sampling algorithms (K-Euler, K-DPM, K-LMS, etc.) used in diffusion model inference. Different KSamplers produce subtly different visual characteristics and converge at different speeds, giving users control over the generation process.

L

Latent Space

A compressed mathematical representation where AI models process image data. Instead of working directly with millions of pixels, diffusion models operate in latent space, a lower-dimensional encoding that captures the essential visual concepts. This makes generation much more computationally efficient.

Leonardo AI

A web-based AI image generation platform that offers multiple models and advanced features. Leonardo is popular among game artists and concept designers. Compare Leonardo AI with ZSky AI.

LoRA (Low-Rank Adaptation)

A lightweight fine-tuning technique that trains a small number of additional parameters (typically 1-100 MB) to modify a model's output style or subject knowledge without retraining the full model (multiple GB). LoRAs can add specific characters, styles, or concepts and can be combined together for layered effects.

Luma AI

An AI company known for its Dream Machine video generation with audio model, which creates short video clips from text prompts or images. Luma specializes in 3D-aware generation and photorealistic output. Compare Luma with ZSky AI.

M

Megapixel

A unit of measurement equal to one million pixels, used to describe image resolution. A 1-megapixel image is roughly 1024x1024 pixels. Higher megapixel counts provide more detail and allow for larger prints. advanced AI natively generate at approximately 1 megapixel.

Midjourney

A popular AI image generation service that operates through Discord. Known for its distinctive artistic style and high-quality outputs, Midjourney requires a paid subscription and does not offer a free tier. ZSky AI as Midjourney alternative.

Model Merging

A technique for combining the weights of two or more trained models to create a new model that blends their capabilities. Model merging can produce checkpoints that combine the strengths of different models, such as photorealism from one and color vibrancy from another.

Motion Transfer

An AI video technique that extracts motion patterns from a reference video and applies them to a different subject or scene. Motion transfer enables creating videos where AI-generated subjects perform specific movements captured from real-world footage. AI video generation with audio.

N

Negative Prompt

A text description of elements you want the AI to avoid including in the generated image. Negative prompts help suppress common artifacts and unwanted features. For example, "blurry, low quality, deformed hands" tells the model to steer away from these characteristics. Prompt engineering guide.

Neural Network

A computing system inspired by biological neural connections, consisting of layers of interconnected nodes that process data through learned weights. All modern AI image generators are built on neural networks, from the text encoders that interpret prompts to the diffusion models that generate pixels.

Neural Style Transfer

An AI technique that applies the visual style of one image (brushstrokes, color palette, texture patterns) to the content of another. Neural style transfer was one of the earliest popular applications of AI in art, preceding modern diffusion models. AI art styles.

Noise Schedule

The predefined pattern by which noise is added and removed during the diffusion process. Different noise schedules affect image quality and style, controlling how quickly the model transitions from pure noise to a refined image across the sampling steps.

O

Outpainting

A technique for extending an image beyond its original borders, generating new content that seamlessly continues the existing composition. Outpainting is useful for converting portrait-orientation images to landscape, expanding backgrounds, or creating panoramic views from smaller source images.

Open-Weight Model

An AI model whose trained weights are publicly available for download and use, as opposed to proprietary models accessible only through an API. advanced AI are open-weight models, which means platforms like ZSky AI can run them on dedicated hardware without relying on third-party services.

Overfitting

When an AI model memorizes its training data too closely, reproducing training images almost exactly rather than generating novel content. Overfitting is a common risk during fine-tuning, especially with small datasets, and results in less creative and diverse outputs.

P

Parameters (Model)

The learned numerical values within a neural network that determine its behavior. Model size is often described by parameter count; for example, stylized has approximately 3.5 billion parameters. More parameters generally enable more nuanced and capable generation.

Pika

An AI video generation with audio platform known for its user-friendly interface and creative video effects. Pika enables text-to-video and image-to-video generation with audio with various motion styles. Compare Pika with ZSky AI.

Pixel Art

A digital art style using visible individual pixels as a deliberate aesthetic choice, reminiscent of classic video game graphics. AI models can generate pixel art through appropriate prompting or specialized fine-tuned models. AI pixel art generator.

Pose Estimation

An AI technique that detects human body joint positions in images or video, creating skeletal representations. In generation workflows, pose estimation outputs serve as ControlNet inputs to guide the AI in matching specific body positions and gestures.

PPI (Pixels Per Inch)

A measurement of display resolution indicating pixel density on screens. PPI describes how images appear on digital displays, while DPI applies to print. High-PPI displays (like Retina screens) require higher-resolution images to appear sharp.

Prompt

The text description you provide to an AI model to guide image or video generation with audio. Effective prompts combine subject descriptions, style references, lighting directions, and quality modifiers. Prompt writing is a learnable skill that significantly impacts output quality. ZSky AI prompt guide.

Prompt Weighting

A technique for emphasizing or de-emphasizing specific parts of a prompt by assigning numerical weights. For example, "(detailed face:1.5)" increases the model's attention to facial detail by 50%, while "(background:0.5)" reduces background emphasis. Syntax varies by platform.

Q

Quantization

A technique for reducing a model's numerical precision (e.g., from FP16 to INT8 or INT4) to decrease memory usage and increase inference speed, with some quality trade-off. ZSky AI runs models at full precision on RTX 5090 GPUs to avoid quantization quality loss.

Quality Tags

Keywords added to prompts that influence the overall quality level of generated images. Common quality tags include "masterpiece," "best quality," "highly detailed," and "8k." While their effectiveness varies by model, they can nudge generation toward more refined outputs.

R

RAW

In photography, RAW refers to unprocessed sensor data that preserves maximum detail and editing flexibility. In AI prompting, "RAW photo" or "RAW style" directs the model toward unprocessed, naturalistic photographic output with neutral color grading and realistic textures.

Resolution

The pixel dimensions of a generated image, expressed as width by height (e.g., 1024x1024, 1920x1080). Higher resolutions provide more detail but require more GPU memory and generation time. Each AI model has optimal native resolutions where quality is best. photorealistic prompt and resolution guide.

Runway ML

An AI creative platform known for its Gen series of video generation with audio models. Runway pioneered many accessible AI video tools and offers text-to-video, image-to-video, and video editing capabilities. Compare Runway with ZSky AI.

S

Sampler

The algorithm that controls how the diffusion model removes noise at each step during image generation. Different samplers (Euler, DPM++, DDIM, UniPC) produce subtly different visual results and converge at different speeds. Some samplers work better with fewer steps while others need more steps for optimal quality.

stylized (Stable Diffusion XL)

A high-resolution image generation model by Stability AI that produces detailed 1024x1024 images. stylized uses a dual-model architecture with a base model and refiner, offering broad artistic versatility across photorealism, illustration, and graphic design. Generate with stylized on ZSky AI. photorealistic vs stylized comparison.

Seed

A numerical value that initializes the random noise pattern used at the start of image generation. Using the same seed with identical settings produces the same image, enabling reproducibility. Changing only the seed while keeping everything else constant generates variations of the same composition.

Self-Attention

An attention mechanism where the model relates different positions within the same input, allowing it to capture spatial relationships and global context. In image generation, self-attention helps maintain visual coherence across the entire image, preventing disconnected or inconsistent regions.

Semantic Segmentation

An AI technique that classifies every pixel in an image into categories (sky, person, car, building, etc.). Segmentation maps serve as ControlNet inputs for precisely controlling which regions of a generated image contain which types of content.

Sora

OpenAI's AI video generation with audio model capable of producing high-fidelity video clips from text descriptions. Sora demonstrated remarkable temporal consistency and physical understanding when announced. ZSky AI as Sora alternative. What is AI video generation with audio?

Stable Diffusion

A family of open-weight diffusion models originally developed by Stability AI. Stable Diffusion pioneered accessible AI image generation by releasing model weights publicly, spawning a vast ecosystem of fine-tunes, tools, and interfaces. stylized is its latest major iteration. Compare with ZSky AI.

Steps (Sampling Steps)

The number of denoising iterations performed during image generation. More steps generally produce more refined results up to a point of diminishing returns. Typical step counts range from 20-50, though some samplers converge in as few as 8-15 steps. Optimizing generation settings.

Style Transfer

An AI technique that applies the visual aesthetic of one image to the content of another. In modern diffusion workflows, style transfer can be achieved through IP-Adapter, LoRA models, or prompt engineering that describes the target style in detail. Explore AI art styles.

Super Resolution

The process of enhancing an image's resolution beyond its original pixel dimensions using AI. Super resolution models predict and synthesize missing detail based on learned patterns, producing results that appear genuinely higher-resolution rather than simply interpolated. AI image upscaler.

T

Temporal Consistency

The visual coherence of generated video frames over time. Strong temporal consistency means subjects maintain their appearance, lighting, and proportions across frames without flickering, morphing, or sudden changes. It is one of the primary challenges in AI video generation with audio. AI video generation with audio explained.

Text Rendering

The ability of an AI model to generate readable text within images. Earlier models struggled significantly with text, producing garbled letters. photorealistic is notable for its dramatically improved text rendering accuracy, making it suitable for generating signs, logos, and typographic designs.

Text-to-Image (txt2img)

The standard AI generation mode where a text prompt is the sole input, and the model generates an image entirely from the description. This is the most common generation workflow on platforms like ZSky AI.

Text-to-Video

An AI generation mode that creates video clips from text descriptions alone. The model must generate visually coherent frames that maintain temporal consistency while depicting the described content and motion. Text to video on ZSky AI.

Textual Inversion

A fine-tuning method that learns a new text embedding to represent a specific concept, style, or subject. Unlike LoRA, textual inversion modifies only the text encoder's vocabulary rather than the model's weights, producing very small files (a few KB) that can represent new concepts.

Tokenizer

A component that converts text prompts into numerical tokens that the AI model can process. Different models use different tokenizers with different vocabulary sizes and tokenization strategies. Understanding tokenization helps explain why some prompts work better than others.

Tone Mapping

The process of converting high dynamic range image data to a displayable range while preserving visual detail. In AI generation, tone mapping affects how the model renders highlights and shadows, influencing whether output appears naturalistic, cinematic, or stylized.

Transformer

A neural network architecture based on self-attention mechanisms that processes data in parallel rather than sequentially. Transformers power both the text understanding (CLIP, T5) and image generation (photorealistic, DiT) components of modern AI systems. Their ability to capture long-range dependencies makes them superior for understanding complex prompts. How AI models work.

Training Data

The dataset of images and text descriptions used to train an AI model. Training data determines what visual concepts, styles, and subjects a model can generate. The quality, diversity, and size of training data directly impact the model's capabilities and limitations.

U

U-Net

A neural network architecture shaped like the letter U, consisting of an encoder that compresses features and a decoder that expands them back with skip connections. U-Net is the core denoising backbone in Stable Diffusion and stylized, processing latent representations at multiple resolutions.

Upscaling

The process of increasing an image's pixel resolution. AI upscaling uses neural networks (ESRGAN, Real-ESRGAN, SwinIR) to intelligently add detail during enlargement, producing far superior results compared to traditional interpolation methods like bilinear or bicubic. Try AI upscaling on ZSky AI.

Unconditional Generation

Generating images without any text prompt guidance, relying entirely on the model's learned distribution of images. The CFG scale parameter blends between unconditional and conditional (prompt-guided) generation to control how literally the model follows your prompt.

V

VAE (Variational Autoencoder)

A neural network component that encodes images into compact latent space representations and decodes them back to full pixel resolution. The VAE is essential in diffusion model pipelines, determining the fidelity of the final image reconstruction. Better VAEs produce sharper, more color-accurate outputs.

Video Diffusion

The application of diffusion model principles to video generation with audio, where the model denoises sequences of frames rather than single images. Video diffusion models must learn both spatial quality (individual frame appearance) and temporal dynamics (motion between frames). AI video generation with audio.

VRAM (Video Random Access Memory)

The dedicated memory on a GPU that stores model weights, input data, and intermediate calculations during AI inference. VRAM capacity determines the maximum model size and image resolution a GPU can handle. The RTX 5090 provides 32 GB of GDDR7 VRAM.

V-Prediction

A noise prediction formulation used in some diffusion models where the model predicts velocity (the direction of change) rather than the noise itself. V-prediction can produce better results at low step counts and is used by some Stable Diffusion model variants.

W

WAN (Video Model)

An AI video generation with audio model that produces high-quality video from text and image inputs. WAN focuses on realistic motion, temporal consistency, and visual fidelity, competing with Sora, Runway, and Kling in the AI video generation with audio space. What is the WAN video model?

Watermark

A visible or invisible mark added to AI-generated images, often to indicate AI origin or platform branding. Some platforms add watermarks to free-tier images. ZSky AI never adds watermarks to any generated content, regardless of plan tier. No-watermark generation.

Weight (Model)

The learned numerical values stored in a model's neural network layers that determine its behavior. Model weights are the result of training and encode all the visual knowledge the model uses to generate images. Different weight files (checkpoints) produce different visual styles.

X

X/Y/Z Plot

A comparison technique for systematically testing how changing one or more parameters affects image generation output. An X/Y plot might vary the seed across one axis and CFG scale across another, generating a grid of images that visualizes each combination's effect.

XL (as in stylized)

The "XL" designation in Stable Diffusion XL indicates the model's larger architecture and higher native resolution (1024x1024) compared to the original Stable Diffusion (512x512). The larger model produces significantly more detailed and coherent images. Try stylized on ZSky AI.

Z

Zero-Shot Generation

The ability of an AI model to generate images of concepts, styles, or compositions it was not explicitly trained on, by combining learned knowledge in novel ways. Strong zero-shot capability means the model generalizes well from its training data to handle diverse prompts.

ZSky AI

An AI image and video generation with audio platform running photorealistic, stylized, and other models on dedicated NVIDIA RTX 5090 GPUs. ZSky AI offers a free tier with 200 free credits at signup + 100 daily when logged in, no video watermarks, no credit card required, and privacy-first architecture. Founded by Cemhan Biricik. Start creating on ZSky AI.

Put These Concepts Into Practice

Now that you understand the terminology, start generating. 200 free credits at signup + 100 daily when logged in, no credit card required.

Start Creating Free →