Limiting the consecutive occurrence
#79
ntk Wrote: Anyone has idea to generate the first half then use rule via OCLplus (GPU) to reverse it to form the missing half?

Disagree.

Pixel Wrote: So, how would sed delete all lines that don't have at least one or more duplicate character any where in the whole line?

This can be done using the command below, add it, slow it, love it xD:
Code:
/\(.\).*\1/!d


Pixel Wrote: I over looked this, well spotted ntk. We should include this.

Code:
/\(DCBA\|EDCB\|FEDC\|GFED\|HGFE\|IHGF\|JIHG\|KJIH\|LKJI\|MLKJ\|NMLK\|ONML\|PONM\|QPON\|RQPO\|SRQP\|TSRQ\|UTSR\|VUTS\|WVUT\|XWVU\|YXWV\|ZYXW\)/d

also it this the right command?

These kind of command slow it the most! That's why I said I can't do it in the first place (When you asked for alphabetical order rule).


Hash-IT Wrote: I am generating Z at the moment and it has been running for 14 hours or so, the output file is only 3.76MB !!!!!!!!

I don’t think SED is fast enough, this needs to be on GPU.

According to this : "AMD dual core 3GHz" It's your CPU : (
I'm on i7 2.4, I can generate 1GB within couple of hours.


Hash-IT Wrote: Another interesting thing to consider is I used ULM’s regular expressions to do some of this filtering on a text file and it does it much faster. I am purely guessing it would be done in 5 hours or so.

I think filtering an existing list may be faster than generating one then pushing it through SED.

I'll check that.

Hash-IT Wrote: I think none of us should seriously get into generating these lists yet until we are all happy that the filter is ok and we have heard from atom. Think about it, it takes many hours to generate possibly days, we will then have to upload / download all the lists and then find somewhere to store them. Unless you have a huge empty drive they will have to be stored zipped which will mean unzipping them every time you need a character then move on to the next.

Agree!


Here are all the commands for Regular Expression:
Code:
(.)\1\1
(.).*\1.*\1
(.).*\1.*(.).*\2
(.).*(.).*\1.*\2
(.).*(.).*\2.*\1
(ABCD|BCDE|CDEF|DEFG|EFGH|FGHI|GHIJ|HIJK|IJKL|JKLM|KLMN|LMNO|MNOP|NOPQ|OPQR|PQRS|QRST|RSTU|STUV|TUVW|UVWX|VWXY|WXYZ)
All the rules so far.


Messages In This Thread
Limiting the consecutive occurrence - by Hash-IT - 05-23-2012, 12:38 AM
RE: Limiting the consecutive occurrence - by ntk - 05-27-2012, 01:28 PM
RE: Limiting the consecutive occurrence - by TAPE - 05-27-2012, 11:36 PM
RE: Limiting the consecutive occurrence - by ntk - 05-28-2012, 09:57 AM
RE: Limiting the consecutive occurrence - by ntk - 05-29-2012, 12:12 AM
RE: Limiting the consecutive occurrence - by ntk - 05-29-2012, 02:19 AM
RE: Limiting the consecutive occurrence - by ntk - 05-29-2012, 04:02 AM
RE: Limiting the consecutive occurrence - by M@LIK - 05-29-2012, 01:47 PM
RE: Limiting the consecutive occurrence - by ntk - 05-30-2012, 10:11 AM
RE: Limiting the consecutive occurrence - by ntk - 06-01-2012, 11:30 PM
RE: Limiting the consecutive occurrence - by ntk - 06-02-2012, 10:48 AM
RE: Limiting the consecutive occurrence - by ntk - 06-03-2012, 12:56 AM
RE: Limiting the consecutive occurrence - by ntk - 06-03-2012, 01:07 PM
RE: Limiting the consecutive occurrence - by ntk - 06-03-2012, 02:14 PM
RE: Limiting the consecutive occurrence - by ntk - 06-05-2012, 02:08 AM
RE: Limiting the consecutive occurrence - by ntk - 06-05-2012, 02:48 AM
RE: Limiting the consecutive occurrence - by aGia - 06-07-2012, 11:26 PM
RE: Limiting the consecutive occurrence - by ntk - 06-08-2012, 12:13 AM
RE: Limiting the consecutive occurrence - by aGia - 06-08-2012, 12:24 AM
RE: Limiting the consecutive occurrence - by ntk - 06-08-2012, 12:29 AM
RE: Limiting the consecutive occurrence - by ntk - 06-08-2012, 01:52 AM
RE: Limiting the consecutive occurrence - by d333 - 09-04-2012, 01:56 PM
RE: Limiting the consecutive occurrence - by atom - 09-04-2012, 02:04 PM