Master Python Yield keyword: Don't Be a Rookie!



Brief Overview of Python yield

In Python, yield is a keyword that turns a function into a generator. Unlike traditional functions that return a value and forget their state, generators maintain their state, making them particularly useful for iterating through large data sets without storing them in memory. When the function containing the yield keyword is called, it doesn’t actually run the code but returns a generator object. The actual code is executed when the generator's __next__() method is invoked, stopping each time it reaches a yield statement, thereby producing a series of values over time, instead of computing them all at once and sending them back like a list.


Importance of Understanding yield

Understanding the yield keyword is crucial for several reasons:

  • Memory Efficiency: Generators are memory-efficient as they yield one item at a time, making them ideal for large-scale data processing tasks.
  • Lazy Evaluation: Generators compute values on-the-fly and thus are useful for representing infinite sequences or streaming data.
  • Code Readability: Using yield can make your code more readable and maintainable by abstracting away the loop where items are consumed.
  • Versatility: The yield keyword is versatile and can be used in various applications, ranging from file I/O to network programming, making it a must-know for anyone aiming to master Python.


Basics for Beginners

1. What Is a Generator?

A generator is a special type of function that returns an iterator. Unlike standard functions, which return a single value and then lose their state, generators allow you to yield a sequence of values, pausing after each one and resuming from the last yield point when called again. This makes generators incredibly memory-efficient and well-suited for tasks that require iterating over large data sets or streams.

1.1 Difference Between Normal Functions and Generators

  • Statefulness: Generators are stateful, meaning they remember their state between calls. Regular functions don't have this property.
  • Return Type: A regular function returns a single value, while a generator returns an iterator which can be used to yield multiple values.
  • Keyword: In regular functions, the return keyword is used. Generators use yield.
  • Resource Utilization: Generators are generally more memory-efficient as they generate items on-the-fly, unlike normal functions that might return a large list, consuming more memory.


2. How Python yield Works

The Python yield keyword is the cornerstone of a generator function. When a generator function is called, it returns an iterator without executing the function code. This iterator can then be looped over to execute the function until it hits a yield, at which point it returns the yielded value and pauses the function's execution. The function can then be resumed from this point the next time the iterator is called.

2.1 Basic Syntax

def my_generator():
    yield "Hello"
    yield "World"

2.2 Simple Examples

Yielding Numbers:

def count_up_to(max):
    count = 1
    while count <= max:
        yield count
        count += 1

counter = count_up_to(5)
for number in counter:



Yielding Strings:

def greet_people():
    yield "Hello, John"
    yield "Hello, Jane"
    yield "Hello, Jim"

greeter = greet_people()
for greeting in greeter:


Hello, John
Hello, Jane
Hello, Jim


3. Comparing return and yield

Understanding the difference between return and Python yield is crucial for grasping how normal functions differ from generator functions in Python. The two keywords primarily differ in the way they handle return values and maintain execution state.

3.1 Return Values

return: When a function with a return statement is called, it returns a specific value (or values wrapped in a data structure) and exits immediately. Once exited, the function does not maintain any state.

def sum_numbers(a, b):
    return a + b

yield: In contrast, a generator function with a Python yield statement returns a generator object. The function can be resumed from its last yield, allowing for multiple values to be yielded sequentially over different calls.

def generate_numbers():
    yield 1
    yield 2
    yield 3

3.2 Execution State

With return: The function's state is not maintained. If you call the function again, it starts fresh, as if it were the first time being invoked. Local variables are reinitialized, and the function logic starts from the beginning.

def count_to_two():
    return 1
    print("This will never print.")

count_to_two()  # Output: "One"
count_to_two()  # Output: "One" (Starts over)

With yield: The function's state is maintained between calls. This means that local variables retain their value, and execution resumes from the statement immediately following the last yield.

def count_to_two():
    yield 1
    yield 2

counter = count_to_two()
next(counter)  # Output: "One"
next(counter)  # Output: "Two" (Resumes where it left off)

Here's a simple way to summarize it:

  • return gives you the final result right away and resets the function's state.
  • yield gives you an intermediate result while preserving the function's state, allowing it to produce a series of results over time.


Intermediate Concepts

1. Why Use Python yield?

Understanding the benefits of using Python yield can help you write more efficient and clean code. The main advantages of Python yield are related to memory efficiency and lazy evaluation.

1.1 Memory Efficiency

A traditional function that returns a list needs to generate all items in the list before it can return, consuming a lot of memory when the list is large. On the other hand, generators only produce one item at a time, consuming memory only when each item is generated and consumed.

Example with return:

def get_range_return(n):
    return [x for x in range(n)]

# This will consume memory for 1 million integers at once.
large_list = get_range_return(10**6)

Example with yield:

def get_range_yield(n):
    for x in range(n):
        yield x

# This will consume memory for one integer at a time.
large_gen = get_range_yield(10**6)

1.2 Lazy Evaluation

Generators are lazy, meaning they generate values on-the-fly. This feature is useful for dealing with large data streams or files that you don't want to load into memory all at once.

Example: Reading Large Files

def read_large_file(file):
    with open(file, 'r') as f:
        for line in f:
            yield line.strip()

# Only one line will be in memory at a time.
for line in read_large_file('large_file.txt'):


2. Understanding Generator Expressions

Generator expressions are a compact way to create generators. They resemble list comprehensions but use parentheses () instead of brackets [].


# List comprehension
[x * 2 for x in range(3)]

# Generator expression
(x * 2 for x in range(3))


  1. Quick Iteration: When you need a quick, one-time iterator.
  2. Inline Usage: Can be used in function arguments where an iterable is expected.

Example 1: Simple Iteration

# List comprehension
squares_list = [x*x for x in range(5)]
# Generator expression
squares_gen = (x*x for x in range(5))

for square in squares_gen:

Example 2: Inline Usage with sum function

# Using list comprehension
print(sum([x*x for x in range(5)]))  # Output: 30

# Using generator expression (More memory-efficient)
print(sum(x*x for x in range(5)))  # Output: 30


3. Using Python yield in Loops

Using Python yield within loops allows you to generate a sequence of values over time. This is particularly useful when dealing with sequences that are computationally expensive to produce.

Example 1: Fibonacci Sequence

def fibonacci(n):
    a, b = 0, 1
    for _ in range(n):
        yield a
        a, b = b, a + b

# Usage
for num in fibonacci(10):

Example 2: Filtering Values

def get_even_numbers(data):
    for number in data:
        if number % 2 == 0:
            yield number

# Usage
data = [1, 2, 3, 4, 5, 6]
even_numbers = get_even_numbers(data)
for num in even_numbers:
    print(num)  # Output: 2, 4, 6

Best Practices:

  • Exception Handling: When using Python yield, always be prepared for the StopIteration exception, which indicates that there are no more items to generate.
  • Documentation: Always document your generator functions properly, indicating what kind of values they will yield.


Advanced Concepts

As you grow more comfortable with Python yield keyword, you'll find that it offers a wide range of advanced features and applications. In this section, we delve into these, including multi-threading, asynchronous programming, delegation to sub-generators, and communication with generators using the send() method.


1. Python yield and Multi-threading

Although Python's Global Interpreter Lock (GIL) can be a bottleneck for multi-threading, generators can still play a role in designing more efficient applications. They are especially useful in I/O-bound or network-bound scenarios where threading can be beneficial.

Example: Parallel File Processing

from threading import Thread

def read_large_file(file, target_list):
    with open(file, 'r') as f:
        for line in f:

# Reading multiple files in parallel
target1, target2 = [], []
t1 = Thread(target=lambda: all(read_large_file('file1.txt', target1)))
t2 = Thread(target=lambda: all(read_large_file('file2.txt', target2)))

Generators, combined with the async and await keywords, can be used in asynchronous programming for non-blocking I/O operations.

Example: Asynchronous Web Scraping

import aiohttp
import asyncio

async def fetch(url):
    async with aiohttp.ClientSession() as session:
        async with session.get(url) as resp:
            return await resp.text()

async def main():
    urls = ['', '']
    tasks = [fetch(url) for url in urls]
    pages = await asyncio.gather(*tasks)

    for url, page in zip(urls, pages):
        print(f'Content from {url[:30]}: {page[:30]}...')
await main()


2. Python yield from Syntax

The yield from syntax allows a generator to delegate part of its operations to another generator. This simplifies the code when one generator function needs to yield values from another.

Example 1: Flatting a Nested List

def flatten(nested_list):
    for item in nested_list:
        if isinstance(item, list):
            yield from flatten(item)
            yield item

nested = [1, [2, [3, 4], 5]]
for x in flatten(nested):
    print(x)  # Output: 1, 2, 3, 4, 5


3. Generator send Method

The send() method allows external code to send data back into a generator function, thereby affecting its behavior.

Example 1: Dynamic Running Average

def running_average():
    total = 0.0
    count = 0
    average = None
    while True:
        term = yield average
        total += term
        count += 1
        average = total / count

# Using send() to update the running average
avg = running_average()
next(avg)  # priming the generator
print(avg.send(5))  # Output: 5.0
print(avg.send(7))  # Output: 6.0
print(avg.send(9))  # Output: 7.0

Best Practices:

  • State Management: Be cautious about the state when using Python yield with multi-threading or asynchronous programming. Ensure your generator functions are designed to handle concurrent access if needed.
  • Error Handling: When using advanced features like Python yield from or send(), always handle exceptions carefully to maintain generator state.


Practical Applications

Understanding the theory behind yield is just the beginning. The true power of Python yield becomes evident when it's put to practical use in a variety of applications. Here are some concrete examples where Python yield shines:

1. Data Streaming

Generators are excellent for building data streaming pipelines. They allow for efficient, on-the-fly processing without requiring all data to be loaded into memory.

Reading Large Files

# A simple example for reading large files line by line
def read_large_file(file_path):
    with open(file_path, 'r') as file:
        for line in file:
            yield line.strip()

# Usage
for line in read_large_file('large_file.txt'):

Web Scraping

Generators can also be useful in web scraping scenarios where data from multiple pages needs to be aggregated.

import requests

def fetch_pages(base_url, num_pages):
    for i in range(1, num_pages + 1):
        url = f"{base_url}/page/{i}"
        response = requests.get(url)
        yield response.content

# Usage
for content in fetch_pages('', 10):

2. Pipeline Architectures

Generators can be used to create modular and memory-efficient data processing pipelines.

Chaining Generators

# Define a series of generator functions
def read_data(filename):
    with open(filename, 'r') as f:
        for line in f:
            yield line.strip()

def filter_data(lines):
    for line in lines:
        if "ERROR" not in line:
            yield line

def transform_data(lines):
    for line in lines:
        yield line.lower()

# Chain the generators together to form a pipeline
pipeline = transform_data(filter_data(read_data('logfile.txt')))

# Process the data
for line in pipeline:

3. Using yield in Web Frameworks


In Django, you can use generators for streaming large QuerySets without loading them entirely into memory.

from django.core.paginator import Paginator

def stream_queryset_as_csv(queryset, batch_size=500):
    paginator = Paginator(queryset, batch_size)
    for page in paginator.page_range:
        for obj in
            yield f"{obj.field1},{obj.field2}\n"


In Flask, you can use generators to stream large files or data streams efficiently.

from flask import Response

def generate_large_csv():
    for row in read_large_file('large_file.txt'):
        yield ','.join(row) + '\n'

def serve_large_csv():
    return Response(generate_large_csv(), content_type='text/csv')


Performance Implications

The Python yield keyword and generators come with their own set of performance implications that can significantly impact the efficiency of your code. Understanding these can help you decide when to use generators and when to opt for alternative data structures or methods.

1. Benchmarking Generators

Benchmarking can give you a good idea about the memory and time efficiency of generators compared to traditional data structures.

Example: Measuring Time and Memory for a List vs. a Generator

import time
import sys

# Measure time and memory for a list
start_time = time.time()
my_list = [x for x in range(1, 1000000)]
end_time = time.time()
print(f"List Time: {end_time - start_time} seconds")
print(f"List Memory: {sys.getsizeof(my_list)} bytes")

# Measure time and memory for a generator
start_time = time.time()
my_gen = (x for x in range(1, 1000000))
end_time = time.time()
print(f"Generator Time: {end_time - start_time} seconds")
print(f"Generator Memory: {sys.getsizeof(my_gen)} bytes")

2. Python yield vs Traditional Data Structures

Generators can be more memory-efficient than lists, but they are not always faster in terms of execution time due to the overhead of generator function calls.

Example: List Comprehension vs. Generator Expression for Summation

# Using list comprehension
start_time = time.time()
print(sum([x * x for x in range(1000000)]))  # Consumes more memory
end_time = time.time()
print(f"List Comprehension Time: {end_time - start_time} seconds")

# Using generator expression
start_time = time.time()
print(sum(x * x for x in range(1000000)))  # More memory-efficient
end_time = time.time()
print(f"Generator Expression Time: {end_time - start_time} seconds")

3. Limitations and Drawbacks

  1. Stateful: Generators are stateful, meaning they can't be reused once exhausted.
  2. Error Handling: Exceptions within generators are tricky since the tracebacks can be misleading.
  3. Debugging: Debugging can be more challenging as compared to traditional loops and data structures.

4. When Not to Use Python yield

While generators are powerful, they are not always the right tool for the job.

  1. Random Access: If you need to access elements by index or require random access, generators are not suitable.
  2. Reusability: Generators are single-use. If you need to traverse the data multiple times, they might not be the best fit.
  3. Complexity: For simple loops that modify a small list or array, using a generator could be overkill.

Example: Inefficient Use of Generator for Random Access

def get_elements(data):
    for item in data:
        yield item

# Creating a generator
elements = get_elements([1, 2, 3, 4, 5])

# This is inefficient if you only need a specific index
for index, element in enumerate(elements):
    if index == 2:
        print(f"Element at index 2 is {element}")



The guide offers a comprehensive look at Python yield keyword, explaining its utility in creating generators. Beginning with the basics, it covers how generators differ from standard functions in that they maintain their state, allowing for memory-efficient data processing. For intermediate users, the guide delves into the benefits of lazy evaluation and how Python yield can be effectively used in loops. Advanced users will find insights into specialized use-cases, such as multi-threading and asynchronous programming, along with syntactical sugar like Python yield from and methods like .send(). Practical applications such as data streaming and web scraping demonstrate yield's real-world utility. Finally, the guide evaluates the performance trade-offs, emphasizing when to opt for generators over traditional data structures. Overall, the guide aims to be a one-stop resource for Python developers of all levels to understand the capabilities and limitations of Python yield.


Related Resources and Further Reading


Views: 36
Deepak Prasad

Deepak Prasad

He is the founder of GoLinuxCloud and brings over a decade of expertise in Linux, Python, Go, Laravel, DevOps, Kubernetes, Git, Shell scripting, OpenShift, AWS, Networking, and Security. With extensive experience, he excels in various domains, from development to DevOps, Networking, and Security, ensuring robust and efficient solutions for diverse projects. You can connect with him on his LinkedIn profile.

Can't find what you're searching for? Let us assist you.

Enter your query below, and we'll provide instant results tailored to your needs.

If my articles on GoLinuxCloud has helped you, kindly consider buying me a coffee as a token of appreciation.

Buy GoLinuxCloud a Coffee

For any other feedbacks or questions you can send mail to

Thank You for your support!!

Leave a Comment