GEN AI - VI - Understanding Diffusion Models, Prompt Engineering, How Text Becomes an Image in MidJourney and Stable Diffusion

GEN AI - VI

content:

21. Understanding Diffusion Models — The Technology Behind MidJourney & Stable Diffusion

22. A Beginner-Friendly Walkthrough — How Text Becomes an Image in MidJourney & Stable Diffusion

23. Key Innovations That Made Modern Diffusion Models Possible

24. How Prompt Engineering Works in MidJourney & Stable Diffusion


Section 21: Understanding Diffusion Models — The Technology Behind MidJourney & Stable Diffusion

Generative AI for images took a massive leap forward with diffusion models, now used in:

  • MidJourney

  • Stable Diffusion

  • Adobe Firefly

  • DALL·E 3

  • RunwayML

Diffusion models produce photorealistic, creative, and stylistically diverse images using a mathematically elegant process:
add noise → learn to remove noise → generate art from pure noise.

This section explains how diffusion models work at a technical yet intuitive level.


21.1 What Are Diffusion Models?

A diffusion model is a neural network trained to reverse a gradual noising process.

Two phases:

  1. Forward diffusion: Add noise to an image step-by-step until only noise remains.

  2. Reverse diffusion: Learn how to remove noise step-by-step until an image appears.

At generation time, the model starts from random noise and performs the reverse process → creating a brand new image.

This makes diffusion models:

  • Stable

  • Highly creative

  • Detail-rich

  • Flexible (style, mood, composition)


21.2 Forward Process — Destroying the Image With Noise

The forward process is simple:

Given:

  • An image ( x_0 )

We add a little Gaussian noise at each step:

[
x_t = x_{t-1} + \sqrt{\beta_t} \cdot \epsilon
]

Where:

  • ( t ) = timestep

  • ( \beta_t ) = noise schedule

  • ( \epsilon ) = noise sampled from a Gaussian

After thousands of steps, the image becomes pure noise.

This process is fixed.
The model does not learn this part.


21.3 Reverse Process — Reconstructing the Image

The reverse process is learned.

Goal:
Given a noisy image, predict and remove noise:

[
\hat{x}{t-1} = f\theta(x_t, t)
]

Where:

  • ( f_\theta ) = the neural network (U-Net)

  • ( x_t ) = noisy image

  • ( t ) = time step

By repeating this thousands of times:

  • Noise slowly disappears

  • Structure emerges

  • Colors form

  • Final image appears

This is why diffusion models are so good at:

  • Fine details

  • Smooth textures

  • Complex lighting

  • High-resolution output


21.4 Why Use Noise? The Beauty of Randomness

Starting from random noise provides:

  • Unlimited creativity

  • True generative ability

  • A different image each time

  • Controllable variation (via random seed)

This is what gives tools like MidJourney their artistic diversity.


21.5 The U-Net Architecture — The Brain of Stable Diffusion

The diffusion model is built on a U-Net, which is shaped like the letter “U.”

It has:

  • Downsampling path → extracts features

  • Bottleneck → captures abstract concepts

  • Upsampling path → rebuilds the image

Skip connections allow the network to:

  • Keep important details

  • Maintain structural consistency

This is what allows diffusion models to produce:

  • Sharp edges

  • Precise shapes

  • High-quality textures


21.6 Text-to-Image Guidance Using Transformers

Diffusion models need a way to understand your text prompt.

They use a text encoder (like CLIP or T5).

Process:

  1. Prompt → becomes token embeddings

  2. Embeddings → guide the noise removal

  3. Model matches image features to prompt meaning

This is how you get:
A cyberpunk cat riding a motorbike
to actually look like a cyberpunk cat riding a motorbike.


21.7 Classifier-Free Guidance (CFG)

CFG allows controlling how strongly the model follows your prompt.

Guidance scale (1–20):

  • Low → creative, unpredictable

  • High → accurate to prompt but less creative

Example:

  • CFG 7.5 is most common

  • CFG 14 produces intense literal interpretations

  • CFG 1–3 produces abstract art


21.8 Latent Space Diffusion — The Core of Stable Diffusion

Stable Diffusion works in latent space, not pixel space.

Benefits:

  • Faster training

  • Faster generation

  • More creativity

  • Higher quality

Process:

  • Image → encode to latent representation

  • Diffusion happens in latent space

  • Latent → decode back to image

This is why Stable Diffusion runs on consumer GPUs.


21.9 Prompt Engineering — Talking to Diffusion Models

Prompts can include:

  • Camera angles (“35mm lens, close-up”)

  • Lighting (“soft lighting, volumetric shadows”)

  • Art styles (“Studio Ghibli style, impressionist oil painting”)

  • Composition (“symmetrical, centered, rule-of-thirds”)

Prompt structure example:

“A futuristic samurai warrior, neon lights, ultra-detailed, ray-tracing, 8K, concept art, by Greg Rutkowski, cinematic lighting”


21.10 Why Diffusion Models Beat GANs

GANs were the previous generation of image generators.

Diffusion models are superior because:

  • More stable training

  • Higher resolution

  • More controllable output

  • Less mode collapse

  • More diverse style generation

GANs guess directly → often unstable
Diffusion removes noise gradually → stable and clean


21.11 Why MidJourney Looks So Artistic

MidJourney uses:

  • Custom-trained diffusion

  • Proprietary upscaling

  • Reinforcement learning

  • Style models

  • Human aesthetic feedback

It prioritizes beauty, composition, and style over realism.

Stable Diffusion prioritizes:

  • Open-source flexibility

  • Fine-grained control

  • Local generation


21.12 Summary

In this section, you learned:

✔ How diffusion models work
✔ Forward vs reverse diffusion
✔ Why noise is used
✔ U-Net architecture
✔ CLIP/T5 text encoding
✔ Latent space diffusion
✔ Prompt engineering
✔ Why diffusion models replaced GANs
✔ How MidJourney achieves artistic output

This understanding forms the backbone of all modern text-to-image AI systems.


Section 22: A Beginner-Friendly Walkthrough — How Text Becomes an Image in MidJourney & Stable Diffusion

Understanding how a single text prompt becomes a stunning piece of artwork is one of the most fascinating parts of Generative AI.
This section breaks down the entire process step-by-step, from the moment you type your prompt to the final rendered image.


22.1 Step-by-Step Overview

Here is the complete pipeline:

  1. User writes a prompt

  2. Prompt is tokenized

  3. Tokens are converted to embeddings

  4. Text encoder (like CLIP/T5) extracts meaning

  5. Diffusion model initializes random noise

  6. The model removes noise step-by-step

  7. Guidance steers the image toward the prompt

  8. Latent image is decoded into pixels

  9. Additional upscaling & beautification (MidJourney)

  10. Final image delivered to user

Let’s explain each stage in detail.


22.2 Step 1 — The User Writes a Prompt

Example:
“A futuristic cyberpunk city with flying cars, neon lights, rainy atmosphere, 8K cinematic style.”

The prompt describes:

  • Objects (city, cars)

  • Setting (cyberpunk, futuristic)

  • Mood (rain, neon)

  • Style (cinematic 8K)

Every part influences the final image.


22.3 Step 2 — Tokenization

The prompt is broken into small units called tokens.

Example:

  • “cyber”, “punk”

  • “city”

  • “neon”

  • “lights”

  • “cinematic”

  • “8”

  • “K”

Tokenization allows the model to:

  • Support multiple languages

  • Handle rare words

  • Interpret creative phrasing


22.4 Step 3 — Tokens → Embeddings

Each token becomes a vector (a list of numbers).

Example:

  • “city” → [0.17, 0.53, -0.21, …]

  • “cyberpunk” → [0.82, -0.11, 0.98, …]

These embeddings capture:

  • Meaning

  • Context

  • Style

  • Semantics


22.5 Step 4 — Text Encoder Extracts Meaning

The text encoder (like CLIP for Stable Diffusion, or a proprietary encoder for MidJourney) converts the entire sentence into a meaningful representation.

The model now “understands”:

  • futuristic

  • cyberpunk

  • rainy

  • neon lighting

  • flying cars

  • cinematic style

This vector influences the entire generation.


22.6 Step 5 — Model Starts With Pure Noise

Diffusion generation starts from random noise, for example:

[noise matrix of shape 512 × 512]

Each random seed produces:

  • Different composition

  • Different colors

  • Different shapes

Even with the same prompt.


22.7 Step 6 — Reverse Diffusion: Removing Noise Step-by-Step

The image begins as noise and slowly becomes clearer.

Each step tries to predict:

  • Edges

  • Shapes

  • Lighting

  • Objects

  • Style

This takes anywhere from 20 to 200 steps, depending on:

  • Model

  • Speed

  • Quality settings


22.8 Step 7 — Guidance From the Prompt

Classifier-Free Guidance (CFG) determines how strongly the prompt influences the output.

Low CFG → more creativity
High CFG → more literal accuracy

Example outputs:

  • CFG 4 → abstract neon chaos

  • CFG 7.5 → balanced cyberpunk city

  • CFG 14 → extremely literal futuristic city

MidJourney typically uses a custom guidance that makes images:

  • More aesthetic

  • More cinematic

  • More stylized


22.9 Step 8 — Latent → Pixel Decoding

Stable Diffusion generates latent images, not raw pixels.

A Variational Autoencoder (VAE) decodes the latent representation into the final image:

  • Adds color

  • Sharpens textures

  • Restores fine details

This is where the image “comes alive.”

Sponsor Key-Word

"This Content Sponsored by SBO Digital Marketing.

Mobile-Based Part-Time Job Opportunity by SBO!

Earn money online by doing simple content publishing and sharing tasks. Here's how:

Job Type: Mobile-based part-time work

Work Involves:

Content publishing

Content sharing on social media

Time Required: As little as 1 hour a day

Earnings: ₹300 or more daily

Requirements:

Active Facebook and Instagram account

Basic knowledge of using mobile and social media

For more details:

WhatsApp your Name and Qualification to 9994104160

a.Online Part Time Jobs from Home

b.Work from Home Jobs Without Investment

c.Freelance Jobs Online for Students

d.Mobile Based Online Jobs

e.Daily Payment Online Jobs

Keyword & Tag: #OnlinePartTimeJob #WorkFromHome #EarnMoneyOnline #PartTimeJob #jobs #jobalerts #withoutinvestmentjob"


22.10 Step 9 — Post-Processing & Upscaling

MidJourney applies:

  • Style filters

  • Lighting adjustments

  • Contrast enhancements

  • Edge sharpening

  • Artistic coherence improvements

Stable Diffusion may use:

  • ESRGAN

  • SwinIR

  • CodeFormer

  • RealESRGAN

These models enhance:

  • Sharpness

  • Resolution

  • Detail


22.11 Step 10 — Final Image Delivered

After all processing, the model outputs the final artwork.

Your single line of text became an AI-generated masterpiece.


22.12 Example End-to-End Breakdown

Prompt:
“A dragon made of galaxies, cosmic wings, glowing nebula background, 4K fantasy illustration.”

The model interprets:

  • Object → dragon

  • Material → galaxies

  • Lighting → glowing nebula

  • Style → fantasy illustration

  • Resolution → 4K

During diffusion:

  • Wings emerge first

  • Galaxy texture forms

  • Colors blend into nebula

  • Dragon outline solidifies

  • Lighting is polished

  • Final image rendered


22.13 Why This Pipeline Works So Well

Because the model:

  • Understands human language

  • Learns visual concepts

  • Reconstructs images step-by-step

  • Combines creativity with math

  • Encodes meaning in vectors

  • Uses large datasets to generalize

This is why:

  • AI can create art in any style

  • Each generation is unique

  • Results improve dramatically with better prompts


22.14 Summary

In this section, you learned how text transforms into images using a diffusion pipeline:

✔ Tokenization
✔ Embedding
✔ Text encoding
✔ Random noise initialization
✔ Reverse diffusion
✔ Prompt guidance
✔ Latent decoding
✔ Upscaling and post-processing
✔ Final rendered image

This is the complete behind-the-scenes journey from prompt to artwork.


Section 23: Key Innovations That Made Modern Diffusion Models Possible

Before diffusion models became the backbone of tools like MidJourney, Stable Diffusion, DALL·E 3, and Adobe Firefly, several major breakthroughs occurred in AI research.
This section explains the core innovations that made high-quality generative image models possible.


23.1 Innovation 1 — Denoising Autoencoders (DAE)

The earliest foundation for diffusion models came from Denoising Autoencoders, introduced by Vincent et al. (2008).

Key idea:

Take an image → add noise → train a model to remove the noise.

This taught neural networks to:

  • understand image structure

  • reconstruct clean images

  • learn robust latent representations

These principles influenced the later design of reverse diffusion.


23.2 Innovation 2 — Score-Based Modeling

Song & Ermon (2019) introduced the idea of score matching:

[
\text{score} = \nabla_x \log p(x)
]

This means:

  • The model learns the direction in which the data is more likely.

  • During generation, the model follows these directions to “walk” from noise to a real image.

Score-based diffusion → later merged with denoising diffusion → created modern diffusion models.


23.3 Innovation 3 — Reversible Stochastic Processes

Diffusion models rely on stochastic differential equations (SDEs).

Two processes:

  1. Forward diffusion (adds noise)

  2. Reverse diffusion (removes noise)

The fact that these processes are mathematically reversible enables:

  • Stable training

  • High-quality generation

  • Clear control over image details


23.4 Innovation 4 — UNet Architecture

The breakthrough architecture for diffusion models is UNet, introduced earlier for segmentation tasks.

Why UNet is perfect for diffusion:

  • Downsampling blocks extract features

  • Bottleneck encodes global structure

  • Upsampling blocks reconstruct images

  • Skip connections retain fine details

This enables:

  • high-resolution images

  • smooth textures

  • sharp edges

  • aesthetic consistency

MidJourney uses a heavily customized UNet-like architecture.

Sponsor Key-Word

"This Content Sponsored by SBO Digital Marketing.

Mobile-Based Part-Time Job Opportunity by SBO!

Earn money online by doing simple content publishing and sharing tasks. Here's how:

Job Type: Mobile-based part-time work

Work Involves:

Content publishing

Content sharing on social media

Time Required: As little as 1 hour a day

Earnings: ₹300 or more daily

Requirements:

Active Facebook and Instagram account

Basic knowledge of using mobile and social media

For more details:

WhatsApp your Name and Qualification to 9994104160

a.Online Part Time Jobs from Home

b.Work from Home Jobs Without Investment

c.Freelance Jobs Online for Students

d.Mobile Based Online Jobs

e.Daily Payment Online Jobs

Keyword & Tag: #OnlinePartTimeJob #WorkFromHome #EarnMoneyOnline #PartTimeJob #jobs #jobalerts #withoutinvestmentjob"


23.5 Innovation 5 — Variational Autoencoders (VAE)

Stable Diffusion introduced the concept of latent diffusion, meaning images are generated in compressed space.

Why VAEs matter:

  • Reduce computation cost by 10×

  • Allow 512×512 or 1024×1024 image generation on consumer GPUs

  • Preserve visual structure while compressing

This innovation made Stable Diffusion possible on normal hardware.


23.6 Innovation 6 — Cross-Attention for Text + Image Alignment

Relating text to image was a major challenge.

Cross-attention solves:

  • How image regions map to words

  • How style words (“cinematic”, “anime”) affect textures

  • How objects relate spatially based on text

It enables:

  • prompt understanding

  • multi-object composition

  • style-following

  • realistic scene arrangement

This is a major reason why modern AI art follows prompts accurately.


23.7 Innovation 7 — CLIP (Contrastive Language–Image Pretraining)

CLIP is one of the biggest breakthroughs.

CLIP learns by:

  • looking at millions of images

  • reading their captions

  • aligning text and image embeddings

Stable Diffusion uses CLIP text encoder to:

  • interpret meaning

  • encode style

  • extract emotional tone

  • understand artistic intent

Without CLIP, AI would fail to follow prompts correctly.


23.8 Innovation 8 — Latent Diffusion (LDM)

Introduced by the COMPVIS team (2022).

Key idea:

Instead of generating images in pixel space, generate them in latent space.

[
z \rightarrow \hat{x}
]

This improves:

  • speed

  • memory usage

  • training efficiency

  • model portability

This innovation enabled:

  • Stable Diffusion

  • Automatic1111

  • RunDiffusion

  • Numerous open-source AI art tools


23.9 Innovation 9 — Classifier-Free Guidance (CFG)

CFG solves the challenge of balancing:

  • creativity

  • prompt accuracy

By mixing two predictions:

  1. With the prompt

  2. Without the prompt

The model can generate images that look:

  • more aligned

  • more detailed

  • more controllable

CFG is a central part of:

  • Stable Diffusion

  • MidJourney

  • DALL·E models


23.10 Innovation 10 — Massive Internet-Scale Datasets

Diffusion models became powerful because they trained on:

  • billions of images

  • text–image pairs

  • diverse styles

  • real-world scenes

  • artistic content

This allowed the models to learn:

  • physics (lighting, perspective)

  • styles (anime, realism, watercolor)

  • object relationships

  • world knowledge

Without giant datasets, modern Gen AI wouldn’t exist.


23.11 Summary of Section 23

Modern diffusion models are possible due to innovations in:

Innovation Contribution
Denoising autoencoders Introduced noise → clean reconstruction
Score-based models Provided mathematical foundation
Stochastic processes Enabled reversible noise removal
UNet Perfect architecture for diffusion
VAE Made diffusion fast & efficient
Cross-attention Linked text to image regions
CLIP Enabled prompt understanding
Latent diffusion Reduced cost & increased quality
CFG Balanced prompt accuracy vs creativity
Large datasets Gave models world-level knowledge

Sponsor Key-Word

"This Content Sponsored by SBO Digital Marketing.

Mobile-Based Part-Time Job Opportunity by SBO!

Earn money online by doing simple content publishing and sharing tasks. Here's how:

Job Type: Mobile-based part-time work

Work Involves:

Content publishing

Content sharing on social media

Time Required: As little as 1 hour a day

Earnings: ₹300 or more daily

Requirements:

Active Facebook and Instagram account

Basic knowledge of using mobile and social media

For more details:

WhatsApp your Name and Qualification to 9994104160

a.Online Part Time Jobs from Home

b.Work from Home Jobs Without Investment

c.Freelance Jobs Online for Students

d.Mobile Based Online Jobs

e.Daily Payment Online Jobs

Keyword & Tag: #OnlinePartTimeJob #WorkFromHome #EarnMoneyOnline #PartTimeJob #jobs #jobalerts #withoutinvestmentjob"


Section 24: How Prompt Engineering Works in MidJourney & Stable Diffusion

Prompt engineering is one of the most important skills in the world of Generative AI.
Even though models like MidJourney, Stable Diffusion, DALL·E, and Adobe Firefly are extremely powerful, the quality of the output almost always depends on the quality of the prompt.

This section explains how prompts are understood, how text is converted into image features, and how you can design highly effective prompts that produce consistent, artistic, and highly controlled results.


24.1 What Is Prompt Engineering?

Prompt Engineering is the science and art of designing text instructions that guide generative models to produce specific outputs.

A prompt provides:

  • Description (what to generate)

  • Context (setting, environment)

  • Style (artistic style, camera type, lighting)

  • Constraints (aspect ratio, quality, details)

  • Modifiers (emotion, mood, realism level)

A well-designed prompt:

  • improves image realism

  • reduces randomness

  • gives finer control

  • produces consistent outputs

MidJourney even calls this the “prompt-to-image language.”


24.2 How Gen AI Models Understand Your Prompts

Modern diffusion models do not understand English like humans.
They break your text into embeddings using:

🔹 CLIP Text Encoder (Stable Diffusion)

🔹 OpenAI Internal Text Encoder (DALL·E)

🔹 Custom Language Models (MidJourney v5/v6)

The text is converted into a vector representation:

[
\text{prompt} \rightarrow \text{embedding} \rightarrow \text{image guidance}
]

These embeddings influence:

  • layout

  • colors

  • shapes

  • scene composition

  • style

  • object interactions


24.3 The Prompt Pipeline: How Text Turns Into Pixels

Here’s the complete internal workflow:

Step 1: Tokenization

The text prompt is broken into tokens (words or subwords).

Step 2: Embedding

CLIP (or another encoder) turns tokens into dense vectors.

Step 3: Cross-Attention

The image generator (UNet) attends to each word and decides:

  • how much each word influences each part of the image

  • which pixels correspond to which concepts

Step 4: Diffusion Sampling

The model starts with noise and repeatedly removes noise while:

  • referencing the text embeddings

  • adjusting the output to match the prompt

Step 5: Upscaling & VAE decoding

Stable Diffusion:

  • generates in latent space

  • decodes with VAE into final RGB image

MidJourney:

  • uses proprietary rendering layers to boost style, lighting, composition


24.4 Types of Prompts (with Examples)

1. Basic Description Prompt

A simple descriptive prompt produces simple output.

Example:
A beautiful sunset over the mountains


2. Detailed Prompt

More details produce more control.

A beautiful golden-hour sunset over the snowy mountains, warm lighting, soft clouds, ultra-realistic


3. Style Prompt

Add styles of artists, eras, or genres.

Cyberpunk street scene, neon lights, rain reflections, in the style of Blade Runner


4. Photography Prompt

Photography-based prompts control:

  • lenses

  • camera types

  • aperture

  • exposure

Example:

Portrait photo of a woman, 85mm lens, f/1.8, soft lighting, shallow depth of field


5. Artistic Movement Prompt

A painting of a samurai warrior, ukiyo-e style, inspired by Hokusai


6. Hybrid Concept Prompt

Combine unrelated concepts.

A dragon made of chocolate flying over Paris, whimsical, Pixar style


24.5 Structure of a High-Quality Prompt

A professional prompt usually has 6 parts:

1. Subject

What is the main object?

2. Medium

Photo? Painting? 3D render? Watercolor?

3. Style

Artist, period, technique.

4. Lighting

Soft light, dramatic shadows, cinematic, neon.

5. Details

Textures, background, environment.

6. Quality Modifiers

— High detail
— 4K / 8K
— Ultra-realistic
— Sharp focus

Sponsor Key-Word

"This Content Sponsored by SBO Digital Marketing.

Mobile-Based Part-Time Job Opportunity by SBO!

Earn money online by doing simple content publishing and sharing tasks. Here's how:

Job Type: Mobile-based part-time work

Work Involves:

Content publishing

Content sharing on social media

Time Required: As little as 1 hour a day

Earnings: ₹300 or more daily

Requirements:

Active Facebook and Instagram account

Basic knowledge of using mobile and social media

For more details:

WhatsApp your Name and Qualification to 9994104160

a.Online Part Time Jobs from Home

b.Work from Home Jobs Without Investment

c.Freelance Jobs Online for Students

d.Mobile Based Online Jobs

e.Daily Payment Online Jobs

Keyword & Tag: #OnlinePartTimeJob #WorkFromHome #EarnMoneyOnline #PartTimeJob #jobs #jobalerts #withoutinvestmentjob"


24.6 MidJourney Prompt Formula (Industry Standard)

The most used formula in professional AI art:

[
\text{(Main Subject)}, \text{medium}, \text{style}, \text{lighting}, \text{details}, \text{mood}, \text{quality}
]

Example:

A futuristic samurai standing in a neon-lit alley, hyper-realistic 3D render, cinematic lighting, rain-soaked streets, dramatic atmosphere, 8K, ultra-detailed


24.7 Stable Diffusion Prompt Formula

Stable Diffusion uses:

  • positive prompts

  • negative prompts

Positive Prompt Example

A majestic lion standing on a cliff, golden hour lighting, ultra-realistic, cinematic, detailed fur

Negative Prompt Example

blurry, distorted face, extra fingers, low resolution, bad anatomy, grainy

Negative prompts heavily improve quality.


24.8 What Happens When Prompts Are Bad?

Poor prompts result in:

  • distorted faces

  • incorrect anatomy

  • unwanted style mixing

  • messy backgrounds

  • incorrect lighting

Example:

A man running

Issues likely:

  • bland image

  • odd proportions

  • wrong movement details

You would need a better description.


24.9 Parameter Controls (MidJourney & Stable Diffusion)

MidJourney

  • --ar (aspect ratio)

  • --chaos (creativity)

  • --style (v4/v5 tone)

  • --s (stylize)

Stable Diffusion

  • CFG scale

  • Sampling steps

  • Sampler type (Euler, DDIM, DPM++ 2M)

  • Seed

  • Negative prompts

These parameters deeply control image accuracy and randomness.


24.10 Prompt Engineering Best Practices

✔ Write clear structure

✔ Use strong nouns and adjectives

✔ Add artistic styles

✔ Use photographic terminology

✔ Use negative prompts

✔ Control lighting

✔ Set the emotion or mood

✔ Use aspect ratios


24.11 Summary

Prompt engineering is a foundational skill for creating stunning AI images.

You learned:

  • how models understand prompts

  • how text becomes image guidance

  • different prompt types

  • formulas for MidJourney & Stable Diffusion

  • best practices

  • parameters for pro-level control

This section prepares you for the next topic.


Sponsor Key-Word

"This Content Sponsored by SBO Digital Marketing.

Mobile-Based Part-Time Job Opportunity by SBO!

Earn money online by doing simple content publishing and sharing tasks. Here's how:

Job Type: Mobile-based part-time work

Work Involves:

Content publishing

Content sharing on social media

Time Required: As little as 1 hour a day

Earnings: ₹300 or more daily

Requirements:

Active Facebook and Instagram account

Basic knowledge of using mobile and social media

For more details:

WhatsApp your Name and Qualification to 9994104160

a.Online Part Time Jobs from Home

b.Work from Home Jobs Without Investment

c.Freelance Jobs Online for Students

d.Mobile Based Online Jobs

e.Daily Payment Online Jobs

Keyword & Tag: #OnlinePartTimeJob #WorkFromHome #EarnMoneyOnline #PartTimeJob #jobs #jobalerts #withoutinvestmentjob"

Comments