1. Observation: Some content I can do up to 2.5x, some 1x
2. Hypothesis: You could probably use PLM perplexity on past content to approximate human encoding
3. Test: Is retention constant after perplexity normalization
www.youtube.com/watch?v=2YYj...
www.youtube.com/watch?v=2YYj...
I don't think you could never make fun of me. I'm like a living parody of myself.
I don't think you could never make fun of me. I'm like a living parody of myself.
1. Basic Stemming
2. Thesaurus/Keyword Expansion
3. Word-Level Learned Latent Semantic Vector Embeddings
4. Chunk Level, Semantic Document Embeddings
And I don't know how to feel about that. I'll keep thinking about it.
1. Basic Stemming
2. Thesaurus/Keyword Expansion
3. Word-Level Learned Latent Semantic Vector Embeddings
4. Chunk Level, Semantic Document Embeddings
And I don't know how to feel about that. I'll keep thinking about it.
seems to strongly support the idea that models can be trained to recognize these patterns. May still be a paper here up for grabs. How do these methods compare in terms of sample-efficiency?
seems to strongly support the idea that models can be trained to recognize these patterns. May still be a paper here up for grabs. How do these methods compare in terms of sample-efficiency?