ringa_lee
Only traversed one by one, thousands of pieces of data are not too much. Similar processing methods are as follows:
# python 2.7 utf-8 from copy import deepcopy dic_category = { u'卫生': [u'扫地', u'拖地', u'吸尘'], u'锻炼': [u'跑步', u'慢跑', u'俯卧撑'], u'自杀': [u'跳楼'] } data = { "Data": [ { "title": u"我要扫地", "id": "1" }, { "title": u"他要跳楼了", "id": "2" }, { "title": u"跑步是有好处的", "id": "3" }, { "title": u"多做俯卧撑", "id": "4" } ] } processed_data = deepcopy(data) # 若考虑内存占用率,直接处理data for dic_ele in processed_data['Data']: dic_ele['category'] = None for str_category, tup_keys in dic_category.iteritems(): if dic_ele['category']: # 不考虑一个title有多种类别的情况 break for str_key in tup_keys: if str_key in dic_ele['title']: dic_ele['category'] = str_category break # display for dic_ele in processed_data['Data']: print '------------' print 'id:', dic_ele['id'] print 'title:', dic_ele['title'].encode('utf-8') print 'category:', dic_ele['category'].encode('utf-8')
I don’t understand what you are talking about at all. Do you mean to only give “sanitation” and find “sweeping the floor” from the text? ? ? That’s really amazing
Only traversed one by one, thousands of pieces of data are not too much. Similar processing methods are as follows:
I don’t understand what you are talking about at all. Do you mean to only give “sanitation” and find “sweeping the floor” from the text? ? ? That’s really amazing