pandas 如何从熊猫数据帧计算 jaccard 相似度

声明:本页面是StackOverFlow热门问题的中英对照翻译,遵循CC BY-SA 4.0协议,如果您需要使用它,必须同样遵循CC BY-SA许可,注明原文地址和作者信息,同时你必须将它归于原作者(不是我):StackOverFlow 原文地址: http://stackoverflow.com/questions/37003272/
Warning: these are provided under cc-by-sa 4.0 license. You are free to use/share it, But you must attribute it to the original authors (not me): StackOverFlow

提示:将鼠标放在中文语句上可以显示对应的英文。显示中英文
时间:2020-09-14 01:10:01  来源:igfitidea点击:

How to compute jaccard similarity from a pandas dataframe

pythonpandasmatrixsimilarity

提问by kitchenprinzessin

I have a dataframe as follows: the shape of the frame is (1510, 1399). The columns represents products, the rows represents the values (0 or 1) assigned by an user for a given product. How can I can compute a jaccard_similarity_score?

我有一个如下的数据框:框的形状是 (1510, 1399)。列代表产品,行代表用户为给定产品分配的值(0 或 1)。如何计算 jaccard_similarity_score?

enter image description here

在此处输入图片说明

I created a placeholder dataframe listing product vs. product

我创建了一个占位符数据框列出产品与产品

data_ibs = pd.DataFrame(index=data_g.columns,columns=data_g.columns)

I am not sure how to iterate though data_ibs to compute similarities.

我不确定如何迭代 data_ibs 来计算相似性。

for i in range(0,len(data_ibs.columns)) :
    # Loop through the columns for each column
    for j in range(0,len(data_ibs.columns)) :
.........

回答by ayhan

Short and vectorized (fast) answer:

简短且矢量化(快速)的答案:

Use 'hamming' from the pairwise distances of scikit learn:

从 scikit learn 的成对距离中使用 'hamming':

from sklearn.metrics.pairwise import pairwise_distances
jac_sim = 1 - pairwise_distances(df.T, metric = "hamming")
# optionally convert it to a DataFrame
jac_sim = pd.DataFrame(jac_sim, index=df.columns, columns=df.columns)


Explanation:

解释:

Assume this is your dataset:

假设这是您的数据集:

import pandas as pd
import numpy as np
np.random.seed(0)
df = pd.DataFrame(np.random.binomial(1, 0.5, size=(100, 5)), columns=list('ABCDE'))
print(df.head())

   A  B  C  D  E
0  1  1  1  1  0
1  1  0  1  1  0
2  1  1  1  1  0
3  0  0  1  1  1
4  1  1  0  1  0

Using sklearn's jaccard_similarity_score, similarity between column A and B is:

使用sklearn的jaccard_similarity_score,A列和B列的相似度为:

from sklearn.metrics import jaccard_similarity_score
print(jaccard_similarity_score(df['A'], df['B']))
0.43

This is the number of rows that have the same value over total number of rows, 100.

这是在总行数 100 中具有相同值的行数。

As far as I know, there is no pairwise version of the jaccard_similarity_score but there are pairwise versions of distances.

据我所知,jaccard_similarity_score 没有成对版本,但有成对版本的距离。

However, SciPy defines Jaccard distanceas follows:

但是,SciPy 将Jaccard 距离定义如下:

Given two vectors, u and v, the Jaccard distance is the proportion of those elements u[i] and v[i] that disagree where at least one of them is non-zero.

给定两个向量 u 和 v,Jaccard 距离是那些不一致的元素 u[i] 和 v[i] 在其中至少一个非零的情况下的比例。

So it excludes the rows where both columns have 0 values. jaccard_similarity_score doesn't. Hamming distance, on the other hand, is inline with the similarity definition:

因此,它排除了两列都为 0 值的行。jaccard_similarity_score 没有。另一方面,汉明距离符合相似性定义:

The proportion of those vector elements between two n-vectors u and v which disagree.

不同意的两个 n 向量 u 和 v 之间的那些向量元素的比例。

So if you want to calculate jaccard_similarity_score, you can use 1 - hamming:

所以如果要计算jaccard_similarity_score,可以使用1-hamming:

from sklearn.metrics.pairwise import pairwise_distances
print(1 - pairwise_distances(df.T, metric = "hamming"))

array([[ 1.  ,  0.43,  0.61,  0.55,  0.46],
       [ 0.43,  1.  ,  0.52,  0.56,  0.49],
       [ 0.61,  0.52,  1.  ,  0.48,  0.53],
       [ 0.55,  0.56,  0.48,  1.  ,  0.49],
       [ 0.46,  0.49,  0.53,  0.49,  1.  ]])

In a DataFrame format:

以数据帧格式:

jac_sim = 1 - pairwise_distances(df.T, metric = "hamming")
jac_sim = pd.DataFrame(jac_sim, index=df.columns, columns=df.columns)
# jac_sim = np.triu(jac_sim) to set the lower diagonal to zero
# jac_sim = np.tril(jac_sim) to set the upper diagonal to zero

      A     B     C     D     E
A  1.00  0.43  0.61  0.55  0.46
B  0.43  1.00  0.52  0.56  0.49
C  0.61  0.52  1.00  0.48  0.53
D  0.55  0.56  0.48  1.00  0.49
E  0.46  0.49  0.53  0.49  1.00

You can do the same by iterating over combinations of columns but it will be much slower.

您可以通过迭代列组合来执行相同的操作,但速度会慢得多。

import itertools
sim_df = pd.DataFrame(np.ones((5, 5)), index=df.columns, columns=df.columns)
for col_pair in itertools.combinations(df.columns, 2):
    sim_df.loc[col_pair] = sim_df.loc[tuple(reversed(col_pair))] = jaccard_similarity_score(df[col_pair[0]], df[col_pair[1]])
print(sim_df)
      A     B     C     D     E
A  1.00  0.43  0.61  0.55  0.46
B  0.43  1.00  0.52  0.56  0.49
C  0.61  0.52  1.00  0.48  0.53
D  0.55  0.56  0.48  1.00  0.49
E  0.46  0.49  0.53  0.49  1.00