<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	
	>
<channel>
	<title>Comments on: How Elasticsearch calculates significant terms</title>
	<atom:link href="http://blog.comperiosearch.com/blog/2015/06/10/how-elasticsearch-calculates-significant-terms/feed/" rel="self" type="application/rss+xml" />
	<link>http://blog.comperiosearch.com/blog/2015/06/10/how-elasticsearch-calculates-significant-terms/</link>
	<description>A blog about Search as THE solution</description>
	<lastBuildDate>Mon, 26 Oct 2015 18:07:52 +0000</lastBuildDate>
		<sy:updatePeriod>hourly</sy:updatePeriod>
		<sy:updateFrequency>1</sy:updateFrequency>
	<generator>https://wordpress.org/?v=3.9.40</generator>
	<item>
		<title>By: André Lynum</title>
		<link>http://blog.comperiosearch.com/blog/2015/06/10/how-elasticsearch-calculates-significant-terms/#comment-19128</link>
		<dc:creator><![CDATA[André Lynum]]></dc:creator>
		<pubDate>Thu, 11 Jun 2015 10:02:39 +0000</pubDate>
		<guid isPermaLink="false">http://blog.comperiosearch.com/?p=3785#comment-19128</guid>
		<description><![CDATA[Ayup, thanks letting us in on the name and mentioning the new sampler aggregation :) That looks quite interesting.

The precision/recall viewpoint is indeed illuminating. I tend to shy away from it since it feels tied to specific task or goal. I&#039;m not really sure I would call bgC3 more precise than Microsoft in this case, although bgC3 probably points to a more limited set of documents. It&#039;s all in the context I guess.

If you can structure your data in a manner that supports your task that is certainly a good idea. Chunking up words I feel is something that has to be done with more care. Pure collocational or n-gram approaches can end up emphasizing turns of phrase without interesting content, you might call them phrase stopwords. I&#039;ld be more inclined to extract meaningful multiword units such as verb complexes or entities.

Doing sentence/paragraph docs incurs indexing costs and seem to me to be a poor mans estimate of the term-frequency. It&#039;s certainly a practical solution but it would nice if there was an option for using the TF in the significant terms aggregation for foreground sets where it is feasible.

Anyhow, thanks for putting cool stuff like this into Elasticsearch and hope you&#039;ll find ways to make significant terms into an even more precise and flexible tool than it is today.

Snakkes!

André]]></description>
		<content:encoded><![CDATA[<p>Ayup, thanks letting us in on the name and mentioning the new sampler aggregation :) That looks quite interesting.</p>
<p>The precision/recall viewpoint is indeed illuminating. I tend to shy away from it since it feels tied to specific task or goal. I&#8217;m not really sure I would call bgC3 more precise than Microsoft in this case, although bgC3 probably points to a more limited set of documents. It&#8217;s all in the context I guess.</p>
<p>If you can structure your data in a manner that supports your task that is certainly a good idea. Chunking up words I feel is something that has to be done with more care. Pure collocational or n-gram approaches can end up emphasizing turns of phrase without interesting content, you might call them phrase stopwords. I&#8217;ld be more inclined to extract meaningful multiword units such as verb complexes or entities.</p>
<p>Doing sentence/paragraph docs incurs indexing costs and seem to me to be a poor mans estimate of the term-frequency. It&#8217;s certainly a practical solution but it would nice if there was an option for using the TF in the significant terms aggregation for foreground sets where it is feasible.</p>
<p>Anyhow, thanks for putting cool stuff like this into Elasticsearch and hope you&#8217;ll find ways to make significant terms into an even more precise and flexible tool than it is today.</p>
<p>Snakkes!</p>
<p>André</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Mark</title>
		<link>http://blog.comperiosearch.com/blog/2015/06/10/how-elasticsearch-calculates-significant-terms/#comment-19127</link>
		<dc:creator><![CDATA[Mark]]></dc:creator>
		<pubDate>Thu, 11 Jun 2015 08:39:20 +0000</pubDate>
		<guid isPermaLink="false">http://blog.comperiosearch.com/?p=3785#comment-19127</guid>
		<description><![CDATA[Nice write up. I plotted some of the other scoring heuristics here: https://twitter.com/elasticmark/status/513320986956292096

For me, the choice of scoring function is essentially a question of emphasis between precision and recall. If I analyse the results of a search for &quot;Bill Gates&quot; do I suggest &quot;Microsoft&quot; or &quot;bgC3&quot; as the most significant terms? The rarer bgc3 is more closely allied (so high precision) but perhaps of less practical use due to poor recall. The inverse is true of Microsoft.

Putting aside the question of precision/recall emphasis,  I found the following features do the most to improve significance suggestions on text:
1) Removal of duplicate/near duplicate text in the results being analysed
2) Discovery of phrases
3) Analysing top-matching samples not all content (see the new &quot;sampler&quot; agg)
4) &quot;Chunking&quot; documents e.g. into sentence-docs if the foreground sample of docs is low in number.
Additionally, text analysis in elasticsearch should ideally not be reliant on field data to avoid memory issues.
These approaches are a little detailed to get into here but are important parts of improving elasticsearch&#039;s significance algos on free text.

The mystery of &quot;JLH&quot; is my daughter&#039;s initials :)
Cheers,
Mark]]></description>
		<content:encoded><![CDATA[<p>Nice write up. I plotted some of the other scoring heuristics here: <a href="https://twitter.com/elasticmark/status/513320986956292096" rel="nofollow">https://twitter.com/elasticmark/status/513320986956292096</a></p>
<p>For me, the choice of scoring function is essentially a question of emphasis between precision and recall. If I analyse the results of a search for &#8220;Bill Gates&#8221; do I suggest &#8220;Microsoft&#8221; or &#8220;bgC3&#8243; as the most significant terms? The rarer bgc3 is more closely allied (so high precision) but perhaps of less practical use due to poor recall. The inverse is true of Microsoft.</p>
<p>Putting aside the question of precision/recall emphasis,  I found the following features do the most to improve significance suggestions on text:<br />
1) Removal of duplicate/near duplicate text in the results being analysed<br />
2) Discovery of phrases<br />
3) Analysing top-matching samples not all content (see the new &#8220;sampler&#8221; agg)<br />
4) &#8220;Chunking&#8221; documents e.g. into sentence-docs if the foreground sample of docs is low in number.<br />
Additionally, text analysis in elasticsearch should ideally not be reliant on field data to avoid memory issues.<br />
These approaches are a little detailed to get into here but are important parts of improving elasticsearch&#8217;s significance algos on free text.</p>
<p>The mystery of &#8220;JLH&#8221; is my daughter&#8217;s initials :)<br />
Cheers,<br />
Mark</p>
]]></content:encoded>
	</item>
</channel>
</rss>
