| <!DOCTYPE html><html lang="en"><head><meta charset="utf-8"><meta name="viewport" content="width=device-width, initial-scale=1.0"><meta name="generator" content="rustdoc"><meta name="description" content="Source of the Rust file `/root/.cargo/git/checkouts/tantivy-65d0bbbddbbd5d02/433372d/src/tokenizer/split_compound_words.rs`."><meta name="keywords" content="rust, rustlang, rust-lang"><title>split_compound_words.rs - source</title><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceSerif4-Regular.ttf.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../FiraSans-Regular.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../FiraSans-Medium.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceCodePro-Regular.ttf.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceSerif4-Bold.ttf.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceCodePro-Semibold.ttf.woff2"><link rel="stylesheet" href="../../../normalize.css"><link rel="stylesheet" href="../../../rustdoc.css" id="mainThemeStyle"><link rel="stylesheet" href="../../../ayu.css" disabled><link rel="stylesheet" href="../../../dark.css" disabled><link rel="stylesheet" href="../../../light.css" id="themeStyle"><script id="default-settings" ></script><script src="../../../storage.js"></script><script defer src="../../../source-script.js"></script><script defer src="../../../source-files.js"></script><script defer src="../../../main.js"></script><noscript><link rel="stylesheet" href="../../../noscript.css"></noscript><link rel="alternate icon" type="image/png" href="../../../favicon-16x16.png"><link rel="alternate icon" type="image/png" href="../../../favicon-32x32.png"><link rel="icon" type="image/svg+xml" href="../../../favicon.svg"></head><body class="rustdoc source"><!--[if lte IE 11]><div class="warning">This old browser is unsupported and will most likely display funky things.</div><![endif]--><nav class="sidebar"><a class="sidebar-logo" href="../../../tantivy/index.html"><div class="logo-container"> |
| <img src="http://fulmicoton.com/tantivy-logo/tantivy-logo.png" alt="logo"></div></a></nav><main><div class="width-limiter"><nav class="sub"><a class="sub-logo-container" href="../../../tantivy/index.html"> |
| <img src="http://fulmicoton.com/tantivy-logo/tantivy-logo.png" alt="logo"></a><form class="search-form"><div class="search-container"><span></span><input class="search-input" name="search" autocomplete="off" spellcheck="false" placeholder="Click or press ‘S’ to search, ‘?’ for more options…" type="search"><div id="help-button" title="help" tabindex="-1"><a href="../../../help.html">?</a></div><div id="settings-menu" tabindex="-1"><a href="../../../settings.html" title="settings"><img width="22" height="22" alt="Change settings" src="../../../wheel.svg"></a></div></div></form></nav><section id="main-content" class="content"><div class="example-wrap"><pre class="src-line-numbers"><span id="1">1</span> |
| <span id="2">2</span> |
| <span id="3">3</span> |
| <span id="4">4</span> |
| <span id="5">5</span> |
| <span id="6">6</span> |
| <span id="7">7</span> |
| <span id="8">8</span> |
| <span id="9">9</span> |
| <span id="10">10</span> |
| <span id="11">11</span> |
| <span id="12">12</span> |
| <span id="13">13</span> |
| <span id="14">14</span> |
| <span id="15">15</span> |
| <span id="16">16</span> |
| <span id="17">17</span> |
| <span id="18">18</span> |
| <span id="19">19</span> |
| <span id="20">20</span> |
| <span id="21">21</span> |
| <span id="22">22</span> |
| <span id="23">23</span> |
| <span id="24">24</span> |
| <span id="25">25</span> |
| <span id="26">26</span> |
| <span id="27">27</span> |
| <span id="28">28</span> |
| <span id="29">29</span> |
| <span id="30">30</span> |
| <span id="31">31</span> |
| <span id="32">32</span> |
| <span id="33">33</span> |
| <span id="34">34</span> |
| <span id="35">35</span> |
| <span id="36">36</span> |
| <span id="37">37</span> |
| <span id="38">38</span> |
| <span id="39">39</span> |
| <span id="40">40</span> |
| <span id="41">41</span> |
| <span id="42">42</span> |
| <span id="43">43</span> |
| <span id="44">44</span> |
| <span id="45">45</span> |
| <span id="46">46</span> |
| <span id="47">47</span> |
| <span id="48">48</span> |
| <span id="49">49</span> |
| <span id="50">50</span> |
| <span id="51">51</span> |
| <span id="52">52</span> |
| <span id="53">53</span> |
| <span id="54">54</span> |
| <span id="55">55</span> |
| <span id="56">56</span> |
| <span id="57">57</span> |
| <span id="58">58</span> |
| <span id="59">59</span> |
| <span id="60">60</span> |
| <span id="61">61</span> |
| <span id="62">62</span> |
| <span id="63">63</span> |
| <span id="64">64</span> |
| <span id="65">65</span> |
| <span id="66">66</span> |
| <span id="67">67</span> |
| <span id="68">68</span> |
| <span id="69">69</span> |
| <span id="70">70</span> |
| <span id="71">71</span> |
| <span id="72">72</span> |
| <span id="73">73</span> |
| <span id="74">74</span> |
| <span id="75">75</span> |
| <span id="76">76</span> |
| <span id="77">77</span> |
| <span id="78">78</span> |
| <span id="79">79</span> |
| <span id="80">80</span> |
| <span id="81">81</span> |
| <span id="82">82</span> |
| <span id="83">83</span> |
| <span id="84">84</span> |
| <span id="85">85</span> |
| <span id="86">86</span> |
| <span id="87">87</span> |
| <span id="88">88</span> |
| <span id="89">89</span> |
| <span id="90">90</span> |
| <span id="91">91</span> |
| <span id="92">92</span> |
| <span id="93">93</span> |
| <span id="94">94</span> |
| <span id="95">95</span> |
| <span id="96">96</span> |
| <span id="97">97</span> |
| <span id="98">98</span> |
| <span id="99">99</span> |
| <span id="100">100</span> |
| <span id="101">101</span> |
| <span id="102">102</span> |
| <span id="103">103</span> |
| <span id="104">104</span> |
| <span id="105">105</span> |
| <span id="106">106</span> |
| <span id="107">107</span> |
| <span id="108">108</span> |
| <span id="109">109</span> |
| <span id="110">110</span> |
| <span id="111">111</span> |
| <span id="112">112</span> |
| <span id="113">113</span> |
| <span id="114">114</span> |
| <span id="115">115</span> |
| <span id="116">116</span> |
| <span id="117">117</span> |
| <span id="118">118</span> |
| <span id="119">119</span> |
| <span id="120">120</span> |
| <span id="121">121</span> |
| <span id="122">122</span> |
| <span id="123">123</span> |
| <span id="124">124</span> |
| <span id="125">125</span> |
| <span id="126">126</span> |
| <span id="127">127</span> |
| <span id="128">128</span> |
| <span id="129">129</span> |
| <span id="130">130</span> |
| <span id="131">131</span> |
| <span id="132">132</span> |
| <span id="133">133</span> |
| <span id="134">134</span> |
| <span id="135">135</span> |
| <span id="136">136</span> |
| <span id="137">137</span> |
| <span id="138">138</span> |
| <span id="139">139</span> |
| <span id="140">140</span> |
| <span id="141">141</span> |
| <span id="142">142</span> |
| <span id="143">143</span> |
| <span id="144">144</span> |
| <span id="145">145</span> |
| <span id="146">146</span> |
| <span id="147">147</span> |
| <span id="148">148</span> |
| <span id="149">149</span> |
| <span id="150">150</span> |
| <span id="151">151</span> |
| <span id="152">152</span> |
| <span id="153">153</span> |
| <span id="154">154</span> |
| <span id="155">155</span> |
| <span id="156">156</span> |
| <span id="157">157</span> |
| <span id="158">158</span> |
| <span id="159">159</span> |
| <span id="160">160</span> |
| <span id="161">161</span> |
| <span id="162">162</span> |
| <span id="163">163</span> |
| <span id="164">164</span> |
| <span id="165">165</span> |
| <span id="166">166</span> |
| <span id="167">167</span> |
| <span id="168">168</span> |
| <span id="169">169</span> |
| <span id="170">170</span> |
| <span id="171">171</span> |
| <span id="172">172</span> |
| <span id="173">173</span> |
| <span id="174">174</span> |
| <span id="175">175</span> |
| <span id="176">176</span> |
| <span id="177">177</span> |
| <span id="178">178</span> |
| <span id="179">179</span> |
| <span id="180">180</span> |
| <span id="181">181</span> |
| <span id="182">182</span> |
| <span id="183">183</span> |
| <span id="184">184</span> |
| <span id="185">185</span> |
| <span id="186">186</span> |
| <span id="187">187</span> |
| <span id="188">188</span> |
| <span id="189">189</span> |
| <span id="190">190</span> |
| <span id="191">191</span> |
| <span id="192">192</span> |
| <span id="193">193</span> |
| <span id="194">194</span> |
| <span id="195">195</span> |
| <span id="196">196</span> |
| <span id="197">197</span> |
| <span id="198">198</span> |
| <span id="199">199</span> |
| <span id="200">200</span> |
| <span id="201">201</span> |
| <span id="202">202</span> |
| <span id="203">203</span> |
| <span id="204">204</span> |
| <span id="205">205</span> |
| <span id="206">206</span> |
| <span id="207">207</span> |
| <span id="208">208</span> |
| <span id="209">209</span> |
| <span id="210">210</span> |
| <span id="211">211</span> |
| <span id="212">212</span> |
| <span id="213">213</span> |
| <span id="214">214</span> |
| <span id="215">215</span> |
| <span id="216">216</span> |
| <span id="217">217</span> |
| <span id="218">218</span> |
| <span id="219">219</span> |
| <span id="220">220</span> |
| <span id="221">221</span> |
| <span id="222">222</span> |
| <span id="223">223</span> |
| <span id="224">224</span> |
| <span id="225">225</span> |
| <span id="226">226</span> |
| <span id="227">227</span> |
| <span id="228">228</span> |
| <span id="229">229</span> |
| <span id="230">230</span> |
| <span id="231">231</span> |
| <span id="232">232</span> |
| <span id="233">233</span> |
| <span id="234">234</span> |
| <span id="235">235</span> |
| <span id="236">236</span> |
| <span id="237">237</span> |
| <span id="238">238</span> |
| <span id="239">239</span> |
| <span id="240">240</span> |
| <span id="241">241</span> |
| <span id="242">242</span> |
| <span id="243">243</span> |
| <span id="244">244</span> |
| <span id="245">245</span> |
| <span id="246">246</span> |
| <span id="247">247</span> |
| <span id="248">248</span> |
| <span id="249">249</span> |
| <span id="250">250</span> |
| <span id="251">251</span> |
| <span id="252">252</span> |
| </pre><pre class="rust"><code><span class="kw">use </span>std::sync::Arc; |
| |
| <span class="kw">use </span>aho_corasick::{AhoCorasick, AhoCorasickBuilder, MatchKind, StateID}; |
| |
| <span class="kw">use super</span>::{BoxTokenStream, Token, TokenFilter, TokenStream}; |
| |
| <span class="doccomment">/// A [`TokenFilter`] which splits compound words into their parts |
| /// based on a given dictionary. |
| /// |
| /// Words only will be split if they can be fully decomposed into |
| /// consecutive matches into the given dictionary. |
| /// |
| /// This is mostly useful to split [compound nouns][compound] common to many |
| /// Germanic languages into their constituents. |
| /// |
| /// # Example |
| /// |
| /// The quality of the dictionary determines the quality of the splits, |
| /// e.g. the missing stem "back" of "backen" implies that "brotbackautomat" |
| /// is not split in the following example. |
| /// |
| /// ```rust |
| /// use tantivy::tokenizer::{SimpleTokenizer, SplitCompoundWords, TextAnalyzer}; |
| /// |
| /// let tokenizer = |
| /// TextAnalyzer::from(SimpleTokenizer).filter(SplitCompoundWords::from_dictionary([ |
| /// "dampf", "schiff", "fahrt", "brot", "backen", "automat", |
| /// ])); |
| /// |
| /// let mut stream = tokenizer.token_stream("dampfschifffahrt"); |
| /// assert_eq!(stream.next().unwrap().text, "dampf"); |
| /// assert_eq!(stream.next().unwrap().text, "schiff"); |
| /// assert_eq!(stream.next().unwrap().text, "fahrt"); |
| /// assert_eq!(stream.next(), None); |
| /// |
| /// let mut stream = tokenizer.token_stream("brotbackautomat"); |
| /// assert_eq!(stream.next().unwrap().text, "brotbackautomat"); |
| /// assert_eq!(stream.next(), None); |
| /// ``` |
| /// |
| /// [compound]: https://en.wikipedia.org/wiki/Compound_(linguistics) |
| </span><span class="attribute">#[derive(Clone)] |
| </span><span class="kw">pub struct </span>SplitCompoundWords<S: StateID> { |
| dict: Arc<AhoCorasick<S>>, |
| } |
| |
| <span class="kw">impl </span>SplitCompoundWords<usize> { |
| <span class="doccomment">/// Create a filter from a given dictionary. |
| /// |
| /// The dictionary will be used to construct an [`AhoCorasick`] automaton |
| /// with reasonable defaults. See [`from_automaton`][Self::from_automaton] if |
| /// more control over its construction is required. |
| </span><span class="kw">pub fn </span>from_dictionary<I, P>(dict: I) -> <span class="self">Self |
| </span><span class="kw">where |
| </span>I: IntoIterator<Item = P>, |
| P: AsRef<[u8]>, |
| { |
| <span class="kw">let </span>dict = AhoCorasickBuilder::new() |
| .match_kind(MatchKind::LeftmostLongest) |
| .build(dict); |
| |
| <span class="self">Self</span>::from_automaton(dict) |
| } |
| } |
| |
| <span class="kw">impl</span><S: StateID> SplitCompoundWords<S> { |
| <span class="doccomment">/// Create a filter from a given automaton. |
| /// |
| /// The automaton should use one of the leftmost-first match kinds |
| /// and it should not be anchored. |
| </span><span class="kw">pub fn </span>from_automaton(dict: AhoCorasick<S>) -> <span class="self">Self </span>{ |
| <span class="self">Self </span>{ |
| dict: Arc::new(dict), |
| } |
| } |
| } |
| |
| <span class="kw">impl</span><S: StateID + Send + Sync + <span class="lifetime">'static</span>> TokenFilter <span class="kw">for </span>SplitCompoundWords<S> { |
| <span class="kw">fn </span>transform<<span class="lifetime">'a</span>>(<span class="kw-2">&</span><span class="self">self</span>, stream: BoxTokenStream<<span class="lifetime">'a</span>>) -> BoxTokenStream<<span class="lifetime">'a</span>> { |
| BoxTokenStream::from(SplitCompoundWordsTokenStream { |
| dict: <span class="self">self</span>.dict.clone(), |
| tail: stream, |
| cuts: Vec::new(), |
| parts: Vec::new(), |
| }) |
| } |
| } |
| |
| <span class="kw">struct </span>SplitCompoundWordsTokenStream<<span class="lifetime">'a</span>, S: StateID> { |
| dict: Arc<AhoCorasick<S>>, |
| tail: BoxTokenStream<<span class="lifetime">'a</span>>, |
| cuts: Vec<usize>, |
| parts: Vec<Token>, |
| } |
| |
| <span class="kw">impl</span><<span class="lifetime">'a</span>, S: StateID> SplitCompoundWordsTokenStream<<span class="lifetime">'a</span>, S> { |
| <span class="comment">// Will use `self.cuts` to fill `self.parts` if `self.tail.token()` |
| // can fully be split into consecutive matches against `self.dict`. |
| </span><span class="kw">fn </span>split(<span class="kw-2">&mut </span><span class="self">self</span>) { |
| <span class="kw">let </span>token = <span class="self">self</span>.tail.token(); |
| <span class="kw">let </span><span class="kw-2">mut </span>text = token.text.as_str(); |
| |
| <span class="self">self</span>.cuts.clear(); |
| <span class="kw">let </span><span class="kw-2">mut </span>pos = <span class="number">0</span>; |
| |
| <span class="kw">for </span>match_ <span class="kw">in </span><span class="self">self</span>.dict.find_iter(text) { |
| <span class="kw">if </span>pos != match_.start() { |
| <span class="kw">break</span>; |
| } |
| |
| <span class="self">self</span>.cuts.push(pos); |
| pos = match_.end(); |
| } |
| |
| <span class="kw">if </span>pos == token.text.len() { |
| <span class="comment">// Fill `self.parts` in reverse order, |
| // so that `self.parts.pop()` yields |
| // the tokens in their original order. |
| </span><span class="kw">for </span>pos <span class="kw">in </span><span class="self">self</span>.cuts.iter().rev() { |
| <span class="kw">let </span>(head, tail) = text.split_at(<span class="kw-2">*</span>pos); |
| |
| text = head; |
| <span class="self">self</span>.parts.push(Token { |
| text: tail.to_owned(), |
| ..<span class="kw-2">*</span>token |
| }); |
| } |
| } |
| } |
| } |
| |
| <span class="kw">impl</span><<span class="lifetime">'a</span>, S: StateID> TokenStream <span class="kw">for </span>SplitCompoundWordsTokenStream<<span class="lifetime">'a</span>, S> { |
| <span class="kw">fn </span>advance(<span class="kw-2">&mut </span><span class="self">self</span>) -> bool { |
| <span class="self">self</span>.parts.pop(); |
| |
| <span class="kw">if </span>!<span class="self">self</span>.parts.is_empty() { |
| <span class="kw">return </span><span class="bool-val">true</span>; |
| } |
| |
| <span class="kw">if </span>!<span class="self">self</span>.tail.advance() { |
| <span class="kw">return </span><span class="bool-val">false</span>; |
| } |
| |
| <span class="comment">// Will yield either `self.parts.last()` or |
| // `self.tail.token()` if it could not be split. |
| </span><span class="self">self</span>.split(); |
| <span class="bool-val">true |
| </span>} |
| |
| <span class="kw">fn </span>token(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="kw-2">&</span>Token { |
| <span class="self">self</span>.parts.last().unwrap_or_else(|| <span class="self">self</span>.tail.token()) |
| } |
| |
| <span class="kw">fn </span>token_mut(<span class="kw-2">&mut </span><span class="self">self</span>) -> <span class="kw-2">&mut </span>Token { |
| <span class="self">self</span>.parts |
| .last_mut() |
| .unwrap_or_else(|| <span class="self">self</span>.tail.token_mut()) |
| } |
| } |
| |
| <span class="attribute">#[cfg(test)] |
| </span><span class="kw">mod </span>tests { |
| <span class="kw">use super</span>::<span class="kw-2">*</span>; |
| <span class="kw">use </span><span class="kw">crate</span>::tokenizer::{SimpleTokenizer, TextAnalyzer}; |
| |
| <span class="attribute">#[test] |
| </span><span class="kw">fn </span>splitting_compound_words_works() { |
| <span class="kw">let </span>tokenizer = TextAnalyzer::from(SimpleTokenizer) |
| .filter(SplitCompoundWords::from_dictionary([<span class="string">"foo"</span>, <span class="string">"bar"</span>])); |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">""</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foo bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foobar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foobarbaz"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foobarbaz"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"baz foobar qux"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"baz"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"qux"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foobar foobar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foobar foo bar foobar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foobazbar foo bar foobar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foobazbar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"foobar qux foobar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"qux"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| |
| { |
| <span class="kw">let </span><span class="kw-2">mut </span>stream = tokenizer.token_stream(<span class="string">"barfoo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"bar"</span>); |
| <span class="macro">assert_eq!</span>(stream.next().unwrap().text, <span class="string">"foo"</span>); |
| <span class="macro">assert_eq!</span>(stream.next(), <span class="prelude-val">None</span>); |
| } |
| } |
| } |
| </code></pre></div> |
| </section></div></main><div id="rustdoc-vars" data-root-path="../../../" data-current-crate="tantivy" data-themes="ayu,dark,light" data-resource-suffix="" data-rustdoc-version="1.66.0-nightly (5c8bff74b 2022-10-21)" ></div></body></html> |