Flatten nested dictionaries, compressing keys

前端 未结 28 2238
遇见更好的自我
遇见更好的自我 2020-11-22 01:16

Suppose you have a dictionary like:

{\'a\': 1,
 \'c\': {\'a\': 2,
       \'b\': {\'x\': 5,
             \'y\' : 10}},
 \'d\': [1, 2, 3]}

Ho

相关标签:
28条回答
  • 2020-11-22 01:34

    There are two big considerations that the original poster needs to consider:

    1. Are there keyspace clobbering issues? For example, {'a_b':{'c':1}, 'a':{'b_c':2}} would result in {'a_b_c':???}. The below solution evades the problem by returning an iterable of pairs.
    2. If performance is an issue, does the key-reducer function (which I hereby refer to as 'join') require access to the entire key-path, or can it just do O(1) work at every node in the tree? If you want to be able to say joinedKey = '_'.join(*keys), that will cost you O(N^2) running time. However if you're willing to say nextKey = previousKey+'_'+thisKey, that gets you O(N) time. The solution below lets you do both (since you could merely concatenate all the keys, then postprocess them).

    (Performance is not likely an issue, but I'll elaborate on the second point in case anyone else cares: In implementing this, there are numerous dangerous choices. If you do this recursively and yield and re-yield, or anything equivalent which touches nodes more than once (which is quite easy to accidentally do), you are doing potentially O(N^2) work rather than O(N). This is because maybe you are calculating a key a then a_1 then a_1_i..., and then calculating a then a_1 then a_1_ii..., but really you shouldn't have to calculate a_1 again. Even if you aren't recalculating it, re-yielding it (a 'level-by-level' approach) is just as bad. A good example is to think about the performance on {1:{1:{1:{1:...(N times)...{1:SOME_LARGE_DICTIONARY_OF_SIZE_N}...}}}})

    Below is a function I wrote flattenDict(d, join=..., lift=...) which can be adapted to many purposes and can do what you want. Sadly it is fairly hard to make a lazy version of this function without incurring the above performance penalties (many python builtins like chain.from_iterable aren't actually efficient, which I only realized after extensive testing of three different versions of this code before settling on this one).

    from collections import Mapping
    from itertools import chain
    from operator import add
    
    _FLAG_FIRST = object()
    
    def flattenDict(d, join=add, lift=lambda x:x):
        results = []
        def visit(subdict, results, partialKey):
            for k,v in subdict.items():
                newKey = lift(k) if partialKey==_FLAG_FIRST else join(partialKey,lift(k))
                if isinstance(v,Mapping):
                    visit(v, results, newKey)
                else:
                    results.append((newKey,v))
        visit(d, results, _FLAG_FIRST)
        return results
    

    To better understand what's going on, below is a diagram for those unfamiliar with reduce(left), otherwise known as "fold left". Sometimes it is drawn with an initial value in place of k0 (not part of the list, passed into the function). Here, J is our join function. We preprocess each kn with lift(k).

                   [k0,k1,...,kN].foldleft(J)
                               /    \
                             ...    kN
                             /
           J(k0,J(k1,J(k2,k3)))
                           /  \
                          /    \
               J(J(k0,k1),k2)   k3
                        /   \
                       /     \
                 J(k0,k1)    k2
                     /  \
                    /    \
                   k0     k1
    

    This is in fact the same as functools.reduce, but where our function does this to all key-paths of the tree.

    >>> reduce(lambda a,b:(a,b), range(5))
    ((((0, 1), 2), 3), 4)
    

    Demonstration (which I'd otherwise put in docstring):

    >>> testData = {
            'a':1,
            'b':2,
            'c':{
                'aa':11,
                'bb':22,
                'cc':{
                    'aaa':111
                }
            }
        }
    from pprint import pprint as pp
    
    >>> pp(dict( flattenDict(testData, lift=lambda x:(x,)) ))
    {('a',): 1,
     ('b',): 2,
     ('c', 'aa'): 11,
     ('c', 'bb'): 22,
     ('c', 'cc', 'aaa'): 111}
    
    >>> pp(dict( flattenDict(testData, join=lambda a,b:a+'_'+b) ))
    {'a': 1, 'b': 2, 'c_aa': 11, 'c_bb': 22, 'c_cc_aaa': 111}    
    
    >>> pp(dict( (v,k) for k,v in flattenDict(testData, lift=hash, join=lambda a,b:hash((a,b))) ))
    {1: 12416037344,
     2: 12544037731,
     11: 5470935132935744593,
     22: 4885734186131977315,
     111: 3461911260025554326}
    

    Performance:

    from functools import reduce
    def makeEvilDict(n):
        return reduce(lambda acc,x:{x:acc}, [{i:0 for i in range(n)}]+range(n))
    
    import timeit
    def time(runnable):
        t0 = timeit.default_timer()
        _ = runnable()
        t1 = timeit.default_timer()
        print('took {:.2f} seconds'.format(t1-t0))
    
    >>> pp(makeEvilDict(8))
    {7: {6: {5: {4: {3: {2: {1: {0: {0: 0,
                                     1: 0,
                                     2: 0,
                                     3: 0,
                                     4: 0,
                                     5: 0,
                                     6: 0,
                                     7: 0}}}}}}}}}
    
    import sys
    sys.setrecursionlimit(1000000)
    
    forget = lambda a,b:''
    
    >>> time(lambda: dict(flattenDict(makeEvilDict(10000), join=forget)) )
    took 0.10 seconds
    >>> time(lambda: dict(flattenDict(makeEvilDict(100000), join=forget)) )
    [1]    12569 segmentation fault  python
    

    ... sigh, don't think that one is my fault...


    [unimportant historical note due to moderation issues]

    Regarding the alleged duplicate of Flatten a dictionary of dictionaries (2 levels deep) of lists in Python:

    That question's solution can be implemented in terms of this one by doing sorted( sum(flatten(...),[]) ). The reverse is not possible: while it is true that the values of flatten(...) can be recovered from the alleged duplicate by mapping a higher-order accumulator, one cannot recover the keys. (edit: Also it turns out that the alleged duplicate owner's question is completely different, in that it only deals with dictionaries exactly 2-level deep, though one of the answers on that page gives a general solution.)

    0 讨论(0)
  • 2020-11-22 01:34

    Davoud's solution is very nice but doesn't give satisfactory results when the nested dict also contains lists of dicts, but his code be adapted for that case:

    def flatten_dict(d):
        items = []
        for k, v in d.items():
            try:
                if (type(v)==type([])): 
                    for l in v: items.extend(flatten_dict(l).items())
                else: 
                    items.extend(flatten_dict(v).items())
            except AttributeError:
                items.append((k, v))
        return dict(items)
    
    0 讨论(0)
  • 2020-11-22 01:35

    How about a functional and performant solution in Python3.5?

    from functools import reduce
    
    
    def _reducer(items, key, val, pref):
        if isinstance(val, dict):
            return {**items, **flatten(val, pref + key)}
        else:
            return {**items, pref + key: val}
    
    def flatten(d, pref=''):
        return(reduce(
            lambda new_d, kv: _reducer(new_d, *kv, pref), 
            d.items(), 
            {}
        ))
    

    This is even more performant:

    def flatten(d, pref=''):
        return(reduce(
            lambda new_d, kv: \
                isinstance(kv[1], dict) and \
                {**new_d, **flatten(kv[1], pref + kv[0])} or \
                {**new_d, pref + kv[0]: kv[1]}, 
            d.items(), 
            {}
        ))
    

    In use:

    my_obj = {'a': 1, 'c': {'a': 2, 'b': {'x': 5, 'y': 10}}, 'd': [1, 2, 3]}
    
    print(flatten(my_obj)) 
    # {'d': [1, 2, 3], 'cby': 10, 'cbx': 5, 'ca': 2, 'a': 1}
    
    0 讨论(0)
  • 2020-11-22 01:36

    Using generators:

    def flat_dic_helper(prepand,d):
        if len(prepand) > 0:
            prepand = prepand + "_"
        for k in d:
            i=d[k]
            if type(i).__name__=='dict':
                r = flat_dic_helper(prepand+k,i)
                for j in r:
                    yield j
            else:
                yield (prepand+k,i)
    
    def flat_dic(d): return dict(flat_dic_helper("",d))
    
    d={'a': 1, 'c': {'a': 2, 'b': {'x': 5, 'y' : 10}}, 'd': [1, 2, 3]}
    print(flat_dic(d))
    
    
    >> {'a': 1, 'c_a': 2, 'c_b_x': 5, 'd': [1, 2, 3], 'c_b_y': 10}
    
    0 讨论(0)
  • 2020-11-22 01:39

    I was thinking of a subclass of UserDict to automagically flat the keys.

    class FlatDict(UserDict):
        def __init__(self, *args, separator='.', **kwargs):
            self.separator = separator
            super().__init__(*args, **kwargs)
    
        def __setitem__(self, key, value):
            if isinstance(value, dict):
                for k1, v1 in FlatDict(value, separator=self.separator).items():
                    super().__setitem__(f"{key}{self.separator}{k1}", v1)
            else:
                super().__setitem__(key, value)
    

    ‌ The advantages it that keys can be added on the fly, or using standard dict instanciation, without surprise:

    >>> fd = FlatDict(
    ...    {
    ...        'person': {
    ...            'sexe': 'male', 
    ...            'name': {
    ...                'first': 'jacques',
    ...                'last': 'dupond'
    ...            }
    ...        }
    ...    }
    ... )
    >>> fd
    {'person.sexe': 'male', 'person.name.first': 'jacques', 'person.name.last': 'dupond'}
    >>> fd['person'] = {'name': {'nickname': 'Bob'}}
    >>> fd
    {'person.sexe': 'male', 'person.name.first': 'jacques', 'person.name.last': 'dupond', 'person.name.nickname': 'Bob'}
    >>> fd['person.name'] = {'civility': 'Dr'}
    >>> fd
    {'person.sexe': 'male', 'person.name.first': 'jacques', 'person.name.last': 'dupond', 'person.name.nickname': 'Bob', 'person.name.civility': 'Dr'}
    
    0 讨论(0)
  • 2020-11-22 01:40

    Using dict.popitem() in straightforward nested-list-like recursion:

    def flatten(d):
        if d == {}:
            return d
        else:
            k,v = d.popitem()
            if (dict != type(v)):
                return {k:v, **flatten(d)}
            else:
                flat_kv = flatten(v)
                for k1 in list(flat_kv.keys()):
                    flat_kv[k + '_' + k1] = flat_kv[k1]
                    del flat_kv[k1]
                return {**flat_kv, **flatten(d)}
    
    0 讨论(0)
提交回复
热议问题