from this SO question

Comment From: hayd

Would this just be on the Series values (i.e. ignore index)?

I had a look in algos and couldn't see anything (np intersect1d seems slow).

OT but a weird thing from that Q is that you can't call Series on a set.

Comment From: cpcloud

yet a frozenset works....werid...i would've though that frozenset was a subclass of set, guess not

Comment From: cpcloud

well it doesn't work...it returns a frozenset...

Comment From: hayd

(I don't see why we enforce this, list is happy to take a set, why shouldn't Series?)

Comment From: cpcloud

because there's no way to map indices to set, they are arbitrary since a set object is unordered

Comment From: hayd

ha! Series let's a lot of stuff drop through.... eg. Series(1)

Comment From: cpcloud

but you're right....list does it so there must be some arbitrary indices assigned

Comment From: cpcloud

i just discovered that Series takes generators! i had no idea.

Comment From: hayd

and that's the workaround, right? pass it to list first... kinda sucks Also isn't a dict similarly unordered (and yet we allow that) :s

Interestingly I thought we used np.fromiter to do that, but apparently it's just list.

Comment From: cpcloud

but dict has keys which are used as the index

Comment From: cpcloud

i guess it's a workaround, but how else would you do it?

Comment From: cpcloud

i made the change....nothing breaks...i'll submit

Comment From: cpcloud

for set construction that is

Comment From: cpcloud

4482

Comment From: jreback

fyi...the Series(1) stuff is getting squash in #3862 (by #3482), its just odd

Comment From: hayd

Cool bananas

Comment From: cpcloud

won't be adding this, so closing

Comment From: ghost

@cpcloud, just making sure your closing this is unrelated to the discussion in #4482, which was only about modifying the series ctor to accept sets, not about adding set operations to series.

Comment From: cpcloud

you're right...that pr was just to disallow frozenset and was only related to the ctor...reopening

Comment From: jreback

@cpcloud what's the status on this?

Comment From: cpcloud

gone by wayside .... i don't really have time to implement this ... but i think we should leave it open ... marking as someday

Comment From: jreback

ok..gr8 thxs

Comment From: jreback

most of these I just push to 0.14.....someday is a box very rarely opened :)

Comment From: makmanalp

Hmm, is this closed now since 0.14 is out?

Comment From: jreback

@makmanalp what are you trying to do?

Comment From: makmanalp

Efficiently calculate which of the rows of a column in df1 also exist in another column in df2 (or perhaps indices instead of columns).

Comment From: jreback

this issue is a bit different than that (see the linked question), did you try isin?

Comment From: hayd

In the original question it looks like the OP wants to ignore the index... in which case they can use the set operations in Index:

pd.Index(s0.values) & pd.Index(s1.values)

Comment From: hayd

Wow, that's really slow, I take that back...

Comment From: jreback

Index keeps things ordered; should'nt do it that way, better to drop into numpy or python, do the set operation and reconstruct.

Comment From: milindsmart

Any update on this? Is this still impossible? I was looking at doing a symmetric difference.

Comment From: h-vetinari

This is something I would have needed several times in the last half year. Using apply is just terribly slow for large data sets (and numpy has fast set implementations like np.intersect1d) - for example, I have code where apply+intersect is 98% of running time.

To recap (since there's a lot of tangential discussion in this thread), I think there is a good case to be made for a .set-accessor, providing access to universal functions operating on sets, like there currently is for .str, .dt. As an example:

import pandas as pd # 0.21.0
import numpy as np # 1.13.3

# this function is just for demonstration purposes
def random_sets(n = 100):
    length = 10
    # strings of random numbers, padded to common length
    s = pd.Series(np.random.randint(0, 10**length-1, (n,), dtype = np.int64)).astype(str).str.zfill(length)
    # split into set of individual numbers and cast back to integers
    return s.map(set).apply(lambda s: set(int(x) for x in s)) 

a = random_sets(5)
# WANTED: a.set.intersect(set([1, 2])) should result in:
a.apply(lambda r: r & set([1, 2])) # intersection of 'a' with {1, 2}, e. g.
# 0    {1, 2}
# 1       {1}
# 2        {}
# 3    {1, 2}
# 4        {}

b = random_sets(5)
# WANTED: a.set.intersect(b) should result in:
pd.concat([a, b], keys = ['a', 'b'], axis = 1).apply(lambda row: row['a'] & row['b'], axis = 1) # intersection of 'a' and 'b' (per row!), e. g.
# 0          {8, 1, 2}
# 1    {1, 3, 5, 7, 9}
# 2       {0, 8, 5, 9}
# 3       {8, 2, 3, 6}
# 4       {8, 9, 3, 6}

Like the .str-methods, it should work with either another Series (including index alignment), or broadcast a 'scalar' set correspondingly. Some important methods I think should be implemented (the function signature tries to indicate the action per row; the names are just suggestions):

(set, set) -> set: a.set.intersect(b) - as above a.set.union(b) - row-wise union of a and b a.set.diff(b) - row-wise set difference of a and b a.set.xor(b) - row-wise symmetric difference of a and b

(set, set) -> bool: a.set.subset(b) - row-wise check if a is subset of b a.set.superset(b) - row-wise check if a is superset of b

(set, obj) -> bool: a.set.contains(c) - row-wise check if a contains c

Comment From: jreback

@h-vetinari sets are not efficiently stored, so this offers only an api benefit, which I have yet to see and interesting use case. You can use Index for these operations individually and that is quite efficient. Series.isin is pretty much .contains. IntervalIndex will work for some of these cases as well.

Comment From: h-vetinari

@jreback, well, I was hoping not just for an API improvement, but some fast cython code to back it up (like for the .str-methods). ;-)

Do I understand you correctly that you propose to work with Series of (short) pd.Indexes? How would you then do something like a.set.intersect(b) (as described above)?

Comment From: jreback

@h-vetinari and you are welcome to contribute things. I the current impl would be quite inefficient and no each way to get around this ATM.

Comment From: chinchillaLiao

import pandas as pd df = pd.DataFrame({'a':{1,2,3}, 'b':{2,3,4}})

Difference operator works between Series.

df['a - b'] = df['a'] - df['b']

Set intersection operactor doesn't work between Series.

df['a & b'] = df['a'] & df['b']

A very slow way to do intersection between Series:

df['a & b'] = df.apply(lambda row: row['a'] & row['b'], axis = 1)

I found it is much more faster to do intersection this way:

df['a & b'] = df['a'] - (df['a'] - df['b'])

I don't know why.

Comment From: h-vetinari

@chinchillaLiao : cool, didn't know set difference worked on Series! It's the only one to work on pandas level though.

But an even better work-around is to go down to the numpy-implementation with .values. In particular, this shouldn't suffer from the speed degradation you're reporting.

(@jreback; my comment half a year ago about a .set-accessor now seems very superfluous - why not just enable the numpy-behaviour directly in pandas?)

df = pd.DataFrame([[{1,2}, {2,3}],[{2,4}, {3, 1}]], columns=['A', 'B'])
df
#         A       B
# 0  {1, 2}  {2, 3}
# 1  {2, 4}  {1, 3}

df['A - B'] = df.A - df.B  # only one that work out of the box in pandas
df['A - B']
# 0       {1}
# 1    {2, 4}
# dtype: object

df['A & B'] = df.A & df.B
# TypeError: unsupported operand type(s) for &: 'set' and 'bool'
df['A & B'] = df.A.values & df.B.values
df['A & B'] 
# 0    {2}
# 1     {}
# Name: A & B, dtype: object

df['A | B'] = df.A | df.B
# TypeError: unsupported operand type(s) for |: 'set' and 'bool'
df['A | B'] = df.A.values | df.B.values
df['A | B'] 
# 0       {1, 2, 3}
# 1    {1, 2, 3, 4}
# Name: A | B, dtype: object

df['A ^ B'] = df.A ^ df.B
# TypeError: unsupported operand type(s) for ^: 'set' and 'bool'
df['A ^ B'] = df.A.values ^ df.B.values
df['A ^ B'] 
# 0          {1, 3}
# 1    {1, 2, 3, 4}
# Name: A ^ B, dtype: object

df
#         A       B   A - B A & B         A | B         A ^ B
# 0  {1, 2}  {2, 3}     {1}   {2}     {1, 2, 3}        {1, 3}
# 1  {2, 4}  {1, 3}  {2, 4}    {}  {1, 2, 3, 4}  {1, 2, 3, 4}

In terms of usability, the really cool thing is that this also works for many-to-one comparisons.

dd = df.A.to_frame()
C = {2, 5}
dd['A - C'] = df.A - C
dd['A & C'] = df.A.values & C
dd['A | C'] = df.A.values | C
dd['A ^ C'] = df.A.values ^ C

dd
#         A A - C A & C      A | C   A ^ C
# 0  {1, 2}   {1}   {2}  {1, 2, 5}  {1, 5}
# 1  {2, 4}   {4}   {2}  {2, 4, 5}  {4, 5}

Comment From: jreback

sets are not first class and actually completely inefficient in a Series

Comment From: h-vetinari

Inefficient as opposed to what? Some situations fundamentally require processing sets.

And even so, why make treating sets harder than it needs to be? I used to think (see my response from December) that this wasn't implemented at all, but since it's in numpy already, why not just expose that functionality on a pandas level? Sure as hell beats writing your own .apply() loops, both in terms of speed and code complexity.

Comment From: jreback

complexity in terms of implementation and code

sure if u wanted to contribute would be great

but it’s not trivial to do in a first class supoorted way

Comment From: hayd

@h-vetinari what is your use case for this? How does this come up?

IMO a nice way to contribute this would be with an extension type in a library. Depending on your use case. If you have a smallish finite super-set you can describe each set as a bitarray (and hence do set operations cheaply).

Note: This is quite different from the original issue: set operations like set(s1) & set(s2)...

Comment From: h-vetinari

@jreback

OK, I'm thinking about contributing that. Since the numpy-methods I showed above are actually not nan-safe,

np.array([{1,2}, np.nan]) | np.array([{2,4}, {3, 1}])
# TypeError: unsupported operand type(s) for |: 'float' and 'set'

I'm back to thinking that a set accessor for Series (not for Index) would be the best. And, since I wouldn't have to write the cython for those methods, I think I can come up with such a wrapper relatively easily.

@hayd Thanks for the link. I've had several use cases over time, e.g. joining email-addresses/telephone numbers when deduplicating user information. But it keeps cropping up. I'm actually really happy I found out about those numpy-methods some days ago. ;-)

Re:

Note: This is quite different from the original issue: set operations like set(s1) & set(s2)...

I've chosen to comment on this issue (rather than opening a new one) due to the title, which imo has a much larger scope. I could easily open a more general issue, if desired.

Comment From: jbrockmendel

Discussed on today's dev call and the consensus was to convert your Series to Index and do setops on those. Closing.