Currently postdoc at @bonsaiseqbioinfo.bsky.social, in Lille.
Investigating patterns (substructures) in structured data (sequences, trees, graphs) of predominantly biological origin.
More at https://fingels.github.io/
bsky.app/profile/npma...
We introduce new ideas to revisit the notion of sampling with window guarantees, also known as minimizers.
A thread:
bsky.app/profile/npma...
There is some bias as the next minimizer after a rescan is more likely to be far as it follows the min over a whole window
There is some bias as the next minimizer after a rescan is more likely to be far as it follows the min over a whole window
--- where E[tau_M] / (M-k+1) is the expected specific density of a random sequence of length M (and then M->infty)
You can derive actually an interval for a finite sequence from this
--- where E[tau_M] / (M-k+1) is the expected specific density of a random sequence of length M (and then M->infty)
You can derive actually an interval for a finite sequence from this
We chose to consider the eps_i as some realizations of an underlying random variable eps of mean 0, but it is only for the proof and not a big deal actually.
We chose to consider the eps_i as some realizations of an underlying random variable eps of mean 0, but it is only for the proof and not a big deal actually.
(and the values E[Z_i] - (w+1)/2 are, equivalently, somewhat around 0)
(and the values E[Z_i] - (w+1)/2 are, equivalently, somewhat around 0)
Take a sequence with n=6, and sample positions 1 and 3.
You get 3 ≠ ((1-0) + (3-1))/2 = 1.5
You need to account for the remaining bits after the last sample. And this implies going to infinity!
Take a sequence with n=6, and sample positions 1 and 3.
You get 3 ≠ ((1-0) + (3-1))/2 = 1.5
You need to account for the remaining bits after the last sample. And this implies going to infinity!