Do iterators save memory in Python?

前端 未结 2 658
花落未央
花落未央 2021-01-12 04:21

I don\'t quite understand how iterators have memory in Python.

>>> l1 = [1, 2, 3, 4, 5, 6]
>>> l2 = [2, 3, 4, 5, 6, 7]
>>> iz = iz         


        
相关标签:
2条回答
  • 2021-01-12 04:53

    Your examples are too simplistic. Consider this:

    nums = [1, 2, 3, 4, 5, 6]
    nums_it = (n for n in nums)
    

    nums_it is a generator that returns all items unmodified from nums. Clearly you do not have any advantage. But consider this:

    squares_it = (n ** 2 for n in nums)
    

    and compare it with:

    squares_lst = [n ** 2 for n in nums]
    

    With squares_it, we are generating the squares of nums on the fly only when requested. With squares_lst, we are generating all of them at once and storing them in a new list.

    So, when you do:

    for n in squares_it:
        print(n)
    

    it's like if you were doing:

    for n in nums:
        print(n ** 2)
    

    But when you do:

    for n in squares_lst:
        print(n)
    

    it's like if you were doing:

    squares_lst = []
    for n in nums:
        squares_lst.append(n ** 2)
    for n in squares_lst:
        print(n)
    

    If you don't need (or don't have) the list nums, then you can save even more space by using:

    squares_it = (n ** 2 for n in xrange(1, 7))
    

    Generators and iterators also provide another significant advantage (which may actually be a disadvantage, depending on the situation): they are evaluated lazily.

    Also, generators and iterators may yield an infinite number of elements. An example is itertools.count() that yields 0, 1, 2, 3, ... without never ending.

    0 讨论(0)
  • 2021-01-12 05:18
    >>> l1 = [1, 2, 3, 4, 5, 6]
    >>> l2 = [2, 3, 4, 5, 6, 7]
    >>> iz = izip(l1, l2)
    

    We still require O(min(l1, l2)) memory as we need to load the lists l1 and l2 in memory.

    With zip you need storage for the two original lists plus the zipped list. With izip you don't store the zipped list.

    Big O notation isn't particularly helpful here if you have to work with a real physical machine instead of some abstract concept of a machine. There's a hidden constant multiplier on your O(n) calculations that could influence the practicality of the code well before n tends to infinity.

    >>> l1 = ( n for n in [1, 2, 3, 4, 5, 6] )
    >>> l2 = ( n for n in [2, 3, 4, 5, 6, 7] )
    >>> iz = izip(l1, l2)
    

    We need to load the lists before converting them into generators, right? This means we'll waste memory. So - what is the point of generators here as well.

    No point to generators here. Any time you see n for n in <expr> without either a more complex expression before the for or an if <expr> filter after it, that's a code smell as you could just have used the original sequence directly. The generators only become useful when you transform the input values into something else or filter them.

    0 讨论(0)
提交回复
热议问题