| <!DOCTYPE html><html lang="en"><head><meta charset="utf-8"><meta name="viewport" content="width=device-width, initial-scale=1.0"><meta name="generator" content="rustdoc"><meta name="description" content="Source of the Rust file `/root/.cargo/git/checkouts/incubator-teaclave-crates-c8106113f74feefc/ede1f68/rusty-machine/src/learning/naive_bayes.rs`."><meta name="keywords" content="rust, rustlang, rust-lang"><title>naive_bayes.rs - source</title><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceSerif4-Regular.ttf.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../FiraSans-Regular.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../FiraSans-Medium.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceCodePro-Regular.ttf.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceSerif4-Bold.ttf.woff2"><link rel="preload" as="font" type="font/woff2" crossorigin href="../../../SourceCodePro-Semibold.ttf.woff2"><link rel="stylesheet" href="../../../normalize.css"><link rel="stylesheet" href="../../../rustdoc.css" id="mainThemeStyle"><link rel="stylesheet" href="../../../ayu.css" disabled><link rel="stylesheet" href="../../../dark.css" disabled><link rel="stylesheet" href="../../../light.css" id="themeStyle"><script id="default-settings" ></script><script src="../../../storage.js"></script><script defer src="../../../source-script.js"></script><script defer src="../../../source-files.js"></script><script defer src="../../../main.js"></script><noscript><link rel="stylesheet" href="../../../noscript.css"></noscript><link rel="alternate icon" type="image/png" href="../../../favicon-16x16.png"><link rel="alternate icon" type="image/png" href="../../../favicon-32x32.png"><link rel="icon" type="image/svg+xml" href="../../../favicon.svg"></head><body class="rustdoc source"><!--[if lte IE 11]><div class="warning">This old browser is unsupported and will most likely display funky things.</div><![endif]--><nav class="sidebar"><a class="sidebar-logo" href="../../../rusty_machine/index.html"><div class="logo-container"><img class="rust-logo" src="../../../rust-logo.svg" alt="logo"></div></a></nav><main><div class="width-limiter"><nav class="sub"><a class="sub-logo-container" href="../../../rusty_machine/index.html"><img class="rust-logo" src="../../../rust-logo.svg" alt="logo"></a><form class="search-form"><div class="search-container"><span></span><input class="search-input" name="search" autocomplete="off" spellcheck="false" placeholder="Click or press ‘S’ to search, ‘?’ for more options…" type="search"><div id="help-button" title="help" tabindex="-1"><a href="../../../help.html">?</a></div><div id="settings-menu" tabindex="-1"><a href="../../../settings.html" title="settings"><img width="22" height="22" alt="Change settings" src="../../../wheel.svg"></a></div></div></form></nav><section id="main-content" class="content"><div class="example-wrap"><pre class="src-line-numbers"><span id="1">1</span> |
| <span id="2">2</span> |
| <span id="3">3</span> |
| <span id="4">4</span> |
| <span id="5">5</span> |
| <span id="6">6</span> |
| <span id="7">7</span> |
| <span id="8">8</span> |
| <span id="9">9</span> |
| <span id="10">10</span> |
| <span id="11">11</span> |
| <span id="12">12</span> |
| <span id="13">13</span> |
| <span id="14">14</span> |
| <span id="15">15</span> |
| <span id="16">16</span> |
| <span id="17">17</span> |
| <span id="18">18</span> |
| <span id="19">19</span> |
| <span id="20">20</span> |
| <span id="21">21</span> |
| <span id="22">22</span> |
| <span id="23">23</span> |
| <span id="24">24</span> |
| <span id="25">25</span> |
| <span id="26">26</span> |
| <span id="27">27</span> |
| <span id="28">28</span> |
| <span id="29">29</span> |
| <span id="30">30</span> |
| <span id="31">31</span> |
| <span id="32">32</span> |
| <span id="33">33</span> |
| <span id="34">34</span> |
| <span id="35">35</span> |
| <span id="36">36</span> |
| <span id="37">37</span> |
| <span id="38">38</span> |
| <span id="39">39</span> |
| <span id="40">40</span> |
| <span id="41">41</span> |
| <span id="42">42</span> |
| <span id="43">43</span> |
| <span id="44">44</span> |
| <span id="45">45</span> |
| <span id="46">46</span> |
| <span id="47">47</span> |
| <span id="48">48</span> |
| <span id="49">49</span> |
| <span id="50">50</span> |
| <span id="51">51</span> |
| <span id="52">52</span> |
| <span id="53">53</span> |
| <span id="54">54</span> |
| <span id="55">55</span> |
| <span id="56">56</span> |
| <span id="57">57</span> |
| <span id="58">58</span> |
| <span id="59">59</span> |
| <span id="60">60</span> |
| <span id="61">61</span> |
| <span id="62">62</span> |
| <span id="63">63</span> |
| <span id="64">64</span> |
| <span id="65">65</span> |
| <span id="66">66</span> |
| <span id="67">67</span> |
| <span id="68">68</span> |
| <span id="69">69</span> |
| <span id="70">70</span> |
| <span id="71">71</span> |
| <span id="72">72</span> |
| <span id="73">73</span> |
| <span id="74">74</span> |
| <span id="75">75</span> |
| <span id="76">76</span> |
| <span id="77">77</span> |
| <span id="78">78</span> |
| <span id="79">79</span> |
| <span id="80">80</span> |
| <span id="81">81</span> |
| <span id="82">82</span> |
| <span id="83">83</span> |
| <span id="84">84</span> |
| <span id="85">85</span> |
| <span id="86">86</span> |
| <span id="87">87</span> |
| <span id="88">88</span> |
| <span id="89">89</span> |
| <span id="90">90</span> |
| <span id="91">91</span> |
| <span id="92">92</span> |
| <span id="93">93</span> |
| <span id="94">94</span> |
| <span id="95">95</span> |
| <span id="96">96</span> |
| <span id="97">97</span> |
| <span id="98">98</span> |
| <span id="99">99</span> |
| <span id="100">100</span> |
| <span id="101">101</span> |
| <span id="102">102</span> |
| <span id="103">103</span> |
| <span id="104">104</span> |
| <span id="105">105</span> |
| <span id="106">106</span> |
| <span id="107">107</span> |
| <span id="108">108</span> |
| <span id="109">109</span> |
| <span id="110">110</span> |
| <span id="111">111</span> |
| <span id="112">112</span> |
| <span id="113">113</span> |
| <span id="114">114</span> |
| <span id="115">115</span> |
| <span id="116">116</span> |
| <span id="117">117</span> |
| <span id="118">118</span> |
| <span id="119">119</span> |
| <span id="120">120</span> |
| <span id="121">121</span> |
| <span id="122">122</span> |
| <span id="123">123</span> |
| <span id="124">124</span> |
| <span id="125">125</span> |
| <span id="126">126</span> |
| <span id="127">127</span> |
| <span id="128">128</span> |
| <span id="129">129</span> |
| <span id="130">130</span> |
| <span id="131">131</span> |
| <span id="132">132</span> |
| <span id="133">133</span> |
| <span id="134">134</span> |
| <span id="135">135</span> |
| <span id="136">136</span> |
| <span id="137">137</span> |
| <span id="138">138</span> |
| <span id="139">139</span> |
| <span id="140">140</span> |
| <span id="141">141</span> |
| <span id="142">142</span> |
| <span id="143">143</span> |
| <span id="144">144</span> |
| <span id="145">145</span> |
| <span id="146">146</span> |
| <span id="147">147</span> |
| <span id="148">148</span> |
| <span id="149">149</span> |
| <span id="150">150</span> |
| <span id="151">151</span> |
| <span id="152">152</span> |
| <span id="153">153</span> |
| <span id="154">154</span> |
| <span id="155">155</span> |
| <span id="156">156</span> |
| <span id="157">157</span> |
| <span id="158">158</span> |
| <span id="159">159</span> |
| <span id="160">160</span> |
| <span id="161">161</span> |
| <span id="162">162</span> |
| <span id="163">163</span> |
| <span id="164">164</span> |
| <span id="165">165</span> |
| <span id="166">166</span> |
| <span id="167">167</span> |
| <span id="168">168</span> |
| <span id="169">169</span> |
| <span id="170">170</span> |
| <span id="171">171</span> |
| <span id="172">172</span> |
| <span id="173">173</span> |
| <span id="174">174</span> |
| <span id="175">175</span> |
| <span id="176">176</span> |
| <span id="177">177</span> |
| <span id="178">178</span> |
| <span id="179">179</span> |
| <span id="180">180</span> |
| <span id="181">181</span> |
| <span id="182">182</span> |
| <span id="183">183</span> |
| <span id="184">184</span> |
| <span id="185">185</span> |
| <span id="186">186</span> |
| <span id="187">187</span> |
| <span id="188">188</span> |
| <span id="189">189</span> |
| <span id="190">190</span> |
| <span id="191">191</span> |
| <span id="192">192</span> |
| <span id="193">193</span> |
| <span id="194">194</span> |
| <span id="195">195</span> |
| <span id="196">196</span> |
| <span id="197">197</span> |
| <span id="198">198</span> |
| <span id="199">199</span> |
| <span id="200">200</span> |
| <span id="201">201</span> |
| <span id="202">202</span> |
| <span id="203">203</span> |
| <span id="204">204</span> |
| <span id="205">205</span> |
| <span id="206">206</span> |
| <span id="207">207</span> |
| <span id="208">208</span> |
| <span id="209">209</span> |
| <span id="210">210</span> |
| <span id="211">211</span> |
| <span id="212">212</span> |
| <span id="213">213</span> |
| <span id="214">214</span> |
| <span id="215">215</span> |
| <span id="216">216</span> |
| <span id="217">217</span> |
| <span id="218">218</span> |
| <span id="219">219</span> |
| <span id="220">220</span> |
| <span id="221">221</span> |
| <span id="222">222</span> |
| <span id="223">223</span> |
| <span id="224">224</span> |
| <span id="225">225</span> |
| <span id="226">226</span> |
| <span id="227">227</span> |
| <span id="228">228</span> |
| <span id="229">229</span> |
| <span id="230">230</span> |
| <span id="231">231</span> |
| <span id="232">232</span> |
| <span id="233">233</span> |
| <span id="234">234</span> |
| <span id="235">235</span> |
| <span id="236">236</span> |
| <span id="237">237</span> |
| <span id="238">238</span> |
| <span id="239">239</span> |
| <span id="240">240</span> |
| <span id="241">241</span> |
| <span id="242">242</span> |
| <span id="243">243</span> |
| <span id="244">244</span> |
| <span id="245">245</span> |
| <span id="246">246</span> |
| <span id="247">247</span> |
| <span id="248">248</span> |
| <span id="249">249</span> |
| <span id="250">250</span> |
| <span id="251">251</span> |
| <span id="252">252</span> |
| <span id="253">253</span> |
| <span id="254">254</span> |
| <span id="255">255</span> |
| <span id="256">256</span> |
| <span id="257">257</span> |
| <span id="258">258</span> |
| <span id="259">259</span> |
| <span id="260">260</span> |
| <span id="261">261</span> |
| <span id="262">262</span> |
| <span id="263">263</span> |
| <span id="264">264</span> |
| <span id="265">265</span> |
| <span id="266">266</span> |
| <span id="267">267</span> |
| <span id="268">268</span> |
| <span id="269">269</span> |
| <span id="270">270</span> |
| <span id="271">271</span> |
| <span id="272">272</span> |
| <span id="273">273</span> |
| <span id="274">274</span> |
| <span id="275">275</span> |
| <span id="276">276</span> |
| <span id="277">277</span> |
| <span id="278">278</span> |
| <span id="279">279</span> |
| <span id="280">280</span> |
| <span id="281">281</span> |
| <span id="282">282</span> |
| <span id="283">283</span> |
| <span id="284">284</span> |
| <span id="285">285</span> |
| <span id="286">286</span> |
| <span id="287">287</span> |
| <span id="288">288</span> |
| <span id="289">289</span> |
| <span id="290">290</span> |
| <span id="291">291</span> |
| <span id="292">292</span> |
| <span id="293">293</span> |
| <span id="294">294</span> |
| <span id="295">295</span> |
| <span id="296">296</span> |
| <span id="297">297</span> |
| <span id="298">298</span> |
| <span id="299">299</span> |
| <span id="300">300</span> |
| <span id="301">301</span> |
| <span id="302">302</span> |
| <span id="303">303</span> |
| <span id="304">304</span> |
| <span id="305">305</span> |
| <span id="306">306</span> |
| <span id="307">307</span> |
| <span id="308">308</span> |
| <span id="309">309</span> |
| <span id="310">310</span> |
| <span id="311">311</span> |
| <span id="312">312</span> |
| <span id="313">313</span> |
| <span id="314">314</span> |
| <span id="315">315</span> |
| <span id="316">316</span> |
| <span id="317">317</span> |
| <span id="318">318</span> |
| <span id="319">319</span> |
| <span id="320">320</span> |
| <span id="321">321</span> |
| <span id="322">322</span> |
| <span id="323">323</span> |
| <span id="324">324</span> |
| <span id="325">325</span> |
| <span id="326">326</span> |
| <span id="327">327</span> |
| <span id="328">328</span> |
| <span id="329">329</span> |
| <span id="330">330</span> |
| <span id="331">331</span> |
| <span id="332">332</span> |
| <span id="333">333</span> |
| <span id="334">334</span> |
| <span id="335">335</span> |
| <span id="336">336</span> |
| <span id="337">337</span> |
| <span id="338">338</span> |
| <span id="339">339</span> |
| <span id="340">340</span> |
| <span id="341">341</span> |
| <span id="342">342</span> |
| <span id="343">343</span> |
| <span id="344">344</span> |
| <span id="345">345</span> |
| <span id="346">346</span> |
| <span id="347">347</span> |
| <span id="348">348</span> |
| <span id="349">349</span> |
| <span id="350">350</span> |
| <span id="351">351</span> |
| <span id="352">352</span> |
| <span id="353">353</span> |
| <span id="354">354</span> |
| <span id="355">355</span> |
| <span id="356">356</span> |
| <span id="357">357</span> |
| <span id="358">358</span> |
| <span id="359">359</span> |
| <span id="360">360</span> |
| <span id="361">361</span> |
| <span id="362">362</span> |
| <span id="363">363</span> |
| <span id="364">364</span> |
| <span id="365">365</span> |
| <span id="366">366</span> |
| <span id="367">367</span> |
| <span id="368">368</span> |
| <span id="369">369</span> |
| <span id="370">370</span> |
| <span id="371">371</span> |
| <span id="372">372</span> |
| <span id="373">373</span> |
| <span id="374">374</span> |
| <span id="375">375</span> |
| <span id="376">376</span> |
| <span id="377">377</span> |
| <span id="378">378</span> |
| <span id="379">379</span> |
| <span id="380">380</span> |
| <span id="381">381</span> |
| <span id="382">382</span> |
| <span id="383">383</span> |
| <span id="384">384</span> |
| <span id="385">385</span> |
| <span id="386">386</span> |
| <span id="387">387</span> |
| <span id="388">388</span> |
| <span id="389">389</span> |
| <span id="390">390</span> |
| <span id="391">391</span> |
| <span id="392">392</span> |
| <span id="393">393</span> |
| <span id="394">394</span> |
| <span id="395">395</span> |
| <span id="396">396</span> |
| <span id="397">397</span> |
| <span id="398">398</span> |
| <span id="399">399</span> |
| <span id="400">400</span> |
| <span id="401">401</span> |
| <span id="402">402</span> |
| <span id="403">403</span> |
| <span id="404">404</span> |
| <span id="405">405</span> |
| <span id="406">406</span> |
| <span id="407">407</span> |
| <span id="408">408</span> |
| <span id="409">409</span> |
| <span id="410">410</span> |
| <span id="411">411</span> |
| <span id="412">412</span> |
| <span id="413">413</span> |
| <span id="414">414</span> |
| <span id="415">415</span> |
| <span id="416">416</span> |
| <span id="417">417</span> |
| <span id="418">418</span> |
| <span id="419">419</span> |
| <span id="420">420</span> |
| <span id="421">421</span> |
| <span id="422">422</span> |
| <span id="423">423</span> |
| <span id="424">424</span> |
| <span id="425">425</span> |
| <span id="426">426</span> |
| <span id="427">427</span> |
| <span id="428">428</span> |
| <span id="429">429</span> |
| <span id="430">430</span> |
| <span id="431">431</span> |
| <span id="432">432</span> |
| <span id="433">433</span> |
| <span id="434">434</span> |
| <span id="435">435</span> |
| <span id="436">436</span> |
| <span id="437">437</span> |
| <span id="438">438</span> |
| <span id="439">439</span> |
| <span id="440">440</span> |
| <span id="441">441</span> |
| <span id="442">442</span> |
| <span id="443">443</span> |
| <span id="444">444</span> |
| <span id="445">445</span> |
| <span id="446">446</span> |
| <span id="447">447</span> |
| <span id="448">448</span> |
| <span id="449">449</span> |
| <span id="450">450</span> |
| <span id="451">451</span> |
| <span id="452">452</span> |
| <span id="453">453</span> |
| <span id="454">454</span> |
| <span id="455">455</span> |
| <span id="456">456</span> |
| <span id="457">457</span> |
| <span id="458">458</span> |
| <span id="459">459</span> |
| <span id="460">460</span> |
| <span id="461">461</span> |
| <span id="462">462</span> |
| <span id="463">463</span> |
| <span id="464">464</span> |
| <span id="465">465</span> |
| <span id="466">466</span> |
| <span id="467">467</span> |
| <span id="468">468</span> |
| <span id="469">469</span> |
| <span id="470">470</span> |
| <span id="471">471</span> |
| <span id="472">472</span> |
| <span id="473">473</span> |
| <span id="474">474</span> |
| <span id="475">475</span> |
| <span id="476">476</span> |
| <span id="477">477</span> |
| <span id="478">478</span> |
| <span id="479">479</span> |
| <span id="480">480</span> |
| <span id="481">481</span> |
| <span id="482">482</span> |
| <span id="483">483</span> |
| <span id="484">484</span> |
| <span id="485">485</span> |
| <span id="486">486</span> |
| <span id="487">487</span> |
| <span id="488">488</span> |
| <span id="489">489</span> |
| <span id="490">490</span> |
| <span id="491">491</span> |
| <span id="492">492</span> |
| <span id="493">493</span> |
| <span id="494">494</span> |
| <span id="495">495</span> |
| <span id="496">496</span> |
| <span id="497">497</span> |
| <span id="498">498</span> |
| <span id="499">499</span> |
| <span id="500">500</span> |
| <span id="501">501</span> |
| <span id="502">502</span> |
| <span id="503">503</span> |
| <span id="504">504</span> |
| <span id="505">505</span> |
| </pre><pre class="rust"><code><span class="doccomment">//! Naive Bayes Classifiers |
| //! |
| //! The classifier supports Gaussian, Bernoulli and Multinomial distributions. |
| //! |
| //! A naive Bayes classifier works by treating the features of each input as independent |
| //! observations. Under this assumption we utilize Bayes' rule to compute the |
| //! probability that each input belongs to a given class. |
| //! |
| //! # Examples |
| //! |
| //! ``` |
| //! use rusty_machine::learning::naive_bayes::{NaiveBayes, Gaussian}; |
| //! use rusty_machine::linalg::Matrix; |
| //! use rusty_machine::learning::SupModel; |
| //! |
| //! let inputs = Matrix::new(6, 2, vec![1.0, 1.1, |
| //! 1.1, 0.9, |
| //! 2.2, 2.3, |
| //! 2.5, 2.7, |
| //! 5.2, 4.3, |
| //! 6.2, 7.3]); |
| //! |
| //! let targets = Matrix::new(6,3, vec![1.0, 0.0, 0.0, |
| //! 1.0, 0.0, 0.0, |
| //! 0.0, 1.0, 0.0, |
| //! 0.0, 1.0, 0.0, |
| //! 0.0, 0.0, 1.0, |
| //! 0.0, 0.0, 1.0]); |
| //! |
| //! // Create a Gaussian Naive Bayes classifier. |
| //! let mut model = NaiveBayes::<Gaussian>::new(); |
| //! |
| //! // Train the model. |
| //! model.train(&inputs, &targets).unwrap(); |
| //! |
| //! // Predict the classes on the input data |
| //! let outputs = model.predict(&inputs).unwrap(); |
| //! |
| //! // Will output the target classes - otherwise our classifier is bad! |
| //! println!("Final outputs --\n{}", outputs); |
| //! ``` |
| |
| </span><span class="kw">use </span>linalg::{Matrix, Axes, BaseMatrix, BaseMatrixMut}; |
| <span class="kw">use </span>learning::{LearningResult, SupModel}; |
| <span class="kw">use </span>learning::error::{Error, ErrorKind}; |
| <span class="kw">use </span>rulinalg::utils; |
| |
| <span class="kw">use </span>std::f64::consts::PI; |
| |
| <span class="doccomment">/// The Naive Bayes model. |
| </span><span class="attribute">#[derive(Debug, Default)] |
| </span><span class="kw">pub struct </span>NaiveBayes<T: Distribution> { |
| distr: <span class="prelude-ty">Option</span><T>, |
| cluster_count: <span class="prelude-ty">Option</span><usize>, |
| class_prior: <span class="prelude-ty">Option</span><Vec<f64>>, |
| class_counts: Vec<usize>, |
| } |
| |
| <span class="kw">impl</span><T: Distribution> NaiveBayes<T> { |
| <span class="doccomment">/// Create a new NaiveBayes model from a given |
| /// distribution. |
| /// |
| /// # Examples |
| /// |
| /// ``` |
| /// use rusty_machine::learning::naive_bayes::{NaiveBayes, Gaussian}; |
| /// |
| /// // Create a new Gaussian Naive Bayes model. |
| /// let _ = NaiveBayes::<Gaussian>::new(); |
| /// ``` |
| </span><span class="kw">pub fn </span>new() -> NaiveBayes<T> { |
| NaiveBayes { |
| distr: <span class="prelude-val">None</span>, |
| cluster_count: <span class="prelude-val">None</span>, |
| class_prior: <span class="prelude-val">None</span>, |
| class_counts: Vec::new(), |
| } |
| } |
| |
| <span class="doccomment">/// Get the cluster count for this model. |
| /// |
| /// Returns an option which is `None` until the model has been trained. |
| </span><span class="kw">pub fn </span>cluster_count(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="prelude-ty">Option</span><<span class="kw-2">&</span>usize> { |
| <span class="self">self</span>.cluster_count.as_ref() |
| } |
| |
| <span class="doccomment">/// Get the class prior distribution for this model. |
| /// |
| /// Returns an option which is `None` until the model has been trained. |
| </span><span class="kw">pub fn </span>class_prior(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="prelude-ty">Option</span><<span class="kw-2">&</span>Vec<f64>> { |
| <span class="self">self</span>.class_prior.as_ref() |
| } |
| |
| <span class="doccomment">/// Get the distribution for this model. |
| /// |
| /// Returns an option which is `None` until the model has been trained. |
| </span><span class="kw">pub fn </span>distr(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="prelude-ty">Option</span><<span class="kw-2">&</span>T> { |
| <span class="self">self</span>.distr.as_ref() |
| } |
| } |
| |
| <span class="doccomment">/// Train and predict from the Naive Bayes model. |
| /// |
| /// The input matrix must be rows made up of features. |
| /// The target matrix should have indicator vectors in each row specifying |
| /// the input class. e.g. [[1,0,0],[0,0,1]] shows class 1 first, then class 3. |
| </span><span class="kw">impl</span><T: Distribution> SupModel<Matrix<f64>, Matrix<f64>> <span class="kw">for </span>NaiveBayes<T> { |
| <span class="doccomment">/// Train the model using inputs and targets. |
| </span><span class="kw">fn </span>train(<span class="kw-2">&mut </span><span class="self">self</span>, inputs: <span class="kw-2">&</span>Matrix<f64>, targets: <span class="kw-2">&</span>Matrix<f64>) -> LearningResult<()> { |
| <span class="self">self</span>.distr = <span class="prelude-val">Some</span>(T::from_model_params(targets.cols(), inputs.cols())); |
| <span class="self">self</span>.update_params(inputs, targets) |
| } |
| |
| <span class="doccomment">/// Predict output from inputs. |
| </span><span class="kw">fn </span>predict(<span class="kw-2">&</span><span class="self">self</span>, inputs: <span class="kw-2">&</span>Matrix<f64>) -> LearningResult<Matrix<f64>> { |
| <span class="kw">let </span>log_probs = <span class="self">self</span>.get_log_probs(inputs)<span class="question-mark">?</span>; |
| <span class="kw">let </span>input_classes = NaiveBayes::<T>::get_classes(log_probs); |
| |
| <span class="kw">if let </span><span class="prelude-val">Some</span>(cluster_count) = <span class="self">self</span>.cluster_count { |
| <span class="kw">let </span><span class="kw-2">mut </span>class_data = Vec::with_capacity(inputs.rows() * cluster_count); |
| |
| <span class="kw">for </span>c <span class="kw">in </span>input_classes { |
| <span class="kw">let </span><span class="kw-2">mut </span>row = <span class="macro">vec!</span>[<span class="number">0f64</span>; cluster_count]; |
| row[c] = <span class="number">1f64</span>; |
| |
| class_data.append(<span class="kw-2">&mut </span>row); |
| } |
| |
| <span class="prelude-val">Ok</span>(Matrix::new(inputs.rows(), cluster_count, class_data)) |
| } <span class="kw">else </span>{ |
| <span class="prelude-val">Err</span>(Error::new(ErrorKind::UntrainedModel, <span class="string">"The model has not been trained."</span>)) |
| } |
| } |
| } |
| |
| <span class="kw">impl</span><T: Distribution> NaiveBayes<T> { |
| <span class="doccomment">/// Get the log-probabilities per class for each input. |
| </span><span class="kw">pub fn </span>get_log_probs(<span class="kw-2">&</span><span class="self">self</span>, inputs: <span class="kw-2">&</span>Matrix<f64>) -> LearningResult<Matrix<f64>> { |
| |
| <span class="kw">if let </span>(<span class="kw-2">&</span><span class="prelude-val">Some</span>(<span class="kw-2">ref </span>distr), <span class="kw-2">&</span><span class="prelude-val">Some</span>(<span class="kw-2">ref </span>prior)) = (<span class="kw-2">&</span><span class="self">self</span>.distr, <span class="kw-2">&</span><span class="self">self</span>.class_prior) { |
| <span class="comment">// Get the joint log likelihood from the distribution |
| </span>distr.joint_log_lik(inputs, prior) |
| } <span class="kw">else </span>{ |
| <span class="prelude-val">Err</span>(Error::new_untrained()) |
| } |
| } |
| |
| <span class="kw">fn </span>update_params(<span class="kw-2">&mut </span><span class="self">self</span>, inputs: <span class="kw-2">&</span>Matrix<f64>, targets: <span class="kw-2">&</span>Matrix<f64>) -> LearningResult<()> { |
| <span class="kw">let </span>class_count = targets.cols(); |
| <span class="kw">let </span>total_data = inputs.rows(); |
| |
| <span class="self">self</span>.class_counts = <span class="macro">vec!</span>[<span class="number">0</span>; class_count]; |
| <span class="kw">let </span><span class="kw-2">mut </span>class_data = <span class="macro">vec!</span>[Vec::new(); class_count]; |
| |
| <span class="kw">for </span>(idx, row) <span class="kw">in </span>targets.row_iter().enumerate() { |
| <span class="comment">// Find the class of this input |
| </span><span class="kw">let </span>class = NaiveBayes::<T>::find_class(row.raw_slice())<span class="question-mark">?</span>; |
| |
| <span class="comment">// Note the class of the input |
| </span>class_data[class].push(idx); |
| <span class="self">self</span>.class_counts[class] += <span class="number">1</span>; |
| } |
| |
| <span class="kw">if let </span><span class="prelude-val">Some</span>(<span class="kw-2">ref mut </span>distr) = <span class="self">self</span>.distr { |
| <span class="kw">for </span>(idx, c) <span class="kw">in </span>class_data.into_iter().enumerate() { |
| <span class="comment">// If this class' vector has not been populated, we can safely |
| // skip this iteration, since the user is clearly not interested |
| // in associating features with this class |
| </span><span class="kw">if </span>c.is_empty() { |
| <span class="kw">continue</span>; |
| } |
| <span class="comment">// Update the parameters within this class |
| </span>distr.update_params(<span class="kw-2">&</span>inputs.select_rows(<span class="kw-2">&</span>c), idx)<span class="question-mark">?</span>; |
| } |
| } |
| |
| <span class="kw">let </span><span class="kw-2">mut </span>class_prior = Vec::with_capacity(class_count); |
| |
| <span class="comment">// Compute the prior as the proportion in each class |
| </span>class_prior.extend(<span class="self">self</span>.class_counts.iter().map(|c| <span class="kw-2">*</span>c <span class="kw">as </span>f64 / total_data <span class="kw">as </span>f64)); |
| |
| <span class="self">self</span>.class_prior = <span class="prelude-val">Some</span>(class_prior); |
| <span class="self">self</span>.cluster_count = <span class="prelude-val">Some</span>(class_count); |
| <span class="prelude-val">Ok</span>(()) |
| } |
| |
| <span class="kw">fn </span>find_class(row: <span class="kw-2">&</span>[f64]) -> LearningResult<usize> { |
| <span class="comment">// Find the `1` entry in the row |
| </span><span class="kw">for </span>(idx, r) <span class="kw">in </span>row.into_iter().enumerate() { |
| <span class="kw">if </span><span class="kw-2">*</span>r == <span class="number">1f64 </span>{ |
| <span class="kw">return </span><span class="prelude-val">Ok</span>(idx); |
| } |
| } |
| |
| <span class="prelude-val">Err</span>(Error::new(ErrorKind::InvalidState, |
| <span class="string">"No class found for entry in targets"</span>)) |
| } |
| |
| <span class="kw">fn </span>get_classes(log_probs: Matrix<f64>) -> Vec<usize> { |
| <span class="kw">let </span><span class="kw-2">mut </span>data_classes = Vec::with_capacity(log_probs.rows()); |
| |
| data_classes.extend(log_probs.row_iter().map(|row| { |
| <span class="comment">// Argmax each class log-probability per input |
| </span><span class="kw">let </span>(class, <span class="kw">_</span>) = utils::argmax(row.raw_slice()); |
| class |
| })); |
| |
| data_classes |
| } |
| } |
| |
| <span class="doccomment">/// Naive Bayes Distribution. |
| </span><span class="kw">pub trait </span>Distribution { |
| <span class="doccomment">/// Initialize the distribution parameters. |
| </span><span class="kw">fn </span>from_model_params(class_count: usize, features: usize) -> <span class="self">Self</span>; |
| |
| <span class="doccomment">/// Updates the distribution parameters. |
| </span><span class="kw">fn </span>update_params(<span class="kw-2">&mut </span><span class="self">self</span>, data: <span class="kw-2">&</span>Matrix<f64>, class: usize) -> LearningResult<()>; |
| |
| <span class="doccomment">/// Compute the joint log likelihood of the data. |
| /// |
| /// Returns a matrix with rows containing the probability that the input lies in each class. |
| </span><span class="kw">fn </span>joint_log_lik(<span class="kw-2">&</span><span class="self">self</span>, |
| data: <span class="kw-2">&</span>Matrix<f64>, |
| class_prior: <span class="kw-2">&</span>[f64]) |
| -> LearningResult<Matrix<f64>>; |
| } |
| |
| <span class="doccomment">/// The Gaussian Naive Bayes model distribution. |
| /// |
| /// Defines: |
| /// |
| /// p(x|C<sub>k</sub>) = ∏<sub>i</sub> N(x<sub>i</sub> ; |
| /// μ<sub>k</sub>, σ<sup>2</sup><sub>k</sub>) |
| </span><span class="attribute">#[derive(Debug)] |
| </span><span class="kw">pub struct </span>Gaussian { |
| theta: Matrix<f64>, |
| sigma: Matrix<f64>, |
| } |
| |
| <span class="kw">impl </span>Gaussian { |
| <span class="doccomment">/// Returns the distribution means. |
| /// |
| /// This is a matrix of class by feature means. |
| </span><span class="kw">pub fn </span>theta(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="kw-2">&</span>Matrix<f64> { |
| <span class="kw-2">&</span><span class="self">self</span>.theta |
| } |
| |
| <span class="doccomment">/// Returns the distribution variances. |
| /// |
| /// This is a matrix of class by feature variances. |
| </span><span class="kw">pub fn </span>sigma(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="kw-2">&</span>Matrix<f64> { |
| <span class="kw-2">&</span><span class="self">self</span>.sigma |
| } |
| } |
| |
| <span class="kw">impl </span>Distribution <span class="kw">for </span>Gaussian { |
| <span class="kw">fn </span>from_model_params(class_count: usize, features: usize) -> Gaussian { |
| Gaussian { |
| theta: Matrix::zeros(class_count, features), |
| sigma: Matrix::zeros(class_count, features), |
| } |
| } |
| |
| <span class="kw">fn </span>update_params(<span class="kw-2">&mut </span><span class="self">self</span>, data: <span class="kw-2">&</span>Matrix<f64>, class: usize) -> LearningResult<()> { |
| <span class="comment">// Compute mean and sample variance |
| </span><span class="kw">let </span>mean = data.mean(Axes::Row).into_vec(); |
| <span class="kw">let </span>var = data.variance(Axes::Row).map_err(|<span class="kw">_</span>| { |
| Error::new(ErrorKind::InvalidData, |
| <span class="string">"Cannot compute variance for Gaussian distribution."</span>) |
| })<span class="question-mark">? |
| </span>.into_vec(); |
| |
| <span class="kw">let </span>features = data.cols(); |
| |
| <span class="kw">for </span>(idx, (m, v)) <span class="kw">in </span>mean.into_iter().zip(var.into_iter()).enumerate() { |
| <span class="self">self</span>.theta.mut_data()[class * features + idx] = m; |
| <span class="self">self</span>.sigma.mut_data()[class * features + idx] = v; |
| } |
| |
| <span class="prelude-val">Ok</span>(()) |
| } |
| |
| <span class="kw">fn </span>joint_log_lik(<span class="kw-2">&</span><span class="self">self</span>, |
| data: <span class="kw-2">&</span>Matrix<f64>, |
| class_prior: <span class="kw-2">&</span>[f64]) |
| -> LearningResult<Matrix<f64>> { |
| <span class="kw">let </span>class_count = class_prior.len(); |
| <span class="kw">let </span><span class="kw-2">mut </span>log_lik = Vec::with_capacity(class_count); |
| |
| <span class="kw">for </span>(i, item) <span class="kw">in </span>class_prior.into_iter().enumerate() { |
| <span class="kw">let </span>joint_i = item.ln(); |
| <span class="kw">let </span>n_ij = -<span class="number">0.5 </span>* (<span class="self">self</span>.sigma.select_rows(<span class="kw-2">&</span>[i]) * <span class="number">2.0 </span>* PI).apply(<span class="kw-2">&</span>|x| x.ln()).sum(); |
| |
| <span class="comment">// NOTE: Here we are copying the row data which is inefficient |
| </span><span class="kw">let </span>r_ij = (data - <span class="self">self</span>.theta.select_rows(<span class="kw-2">&</span><span class="macro">vec!</span>[i; data.rows()])) |
| .apply(<span class="kw-2">&</span>|x| x * x) |
| .elediv(<span class="kw-2">&</span><span class="self">self</span>.sigma.select_rows(<span class="kw-2">&</span><span class="macro">vec!</span>[i; data.rows()])) |
| .sum_cols(); |
| |
| <span class="kw">let </span>res = (-r_ij * <span class="number">0.5</span>) + n_ij; |
| |
| log_lik.append(<span class="kw-2">&mut </span>(res + joint_i).into_vec()); |
| } |
| |
| <span class="prelude-val">Ok</span>(Matrix::new(class_count, data.rows(), log_lik).transpose()) |
| } |
| } |
| |
| <span class="doccomment">/// The Bernoulli Naive Bayes model distribution. |
| /// |
| /// Defines: |
| /// |
| /// p(x|C<sub>k</sub>) = ∏<sub>i</sub> p<sub>k</sub><sup>x<sub>i</sub></sup> |
| /// (1-p)<sub>k</sub><sup>1-x<sub>i</sub></sup> |
| </span><span class="attribute">#[derive(Debug)] |
| </span><span class="kw">pub struct </span>Bernoulli { |
| log_probs: Matrix<f64>, |
| pseudo_count: f64, |
| } |
| |
| <span class="kw">impl </span>Bernoulli { |
| <span class="doccomment">/// The log probability matrix. |
| /// |
| /// A matrix of class by feature model log-probabilities. |
| </span><span class="kw">pub fn </span>log_probs(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="kw-2">&</span>Matrix<f64> { |
| <span class="kw-2">&</span><span class="self">self</span>.log_probs |
| } |
| } |
| |
| <span class="kw">impl </span>Distribution <span class="kw">for </span>Bernoulli { |
| <span class="kw">fn </span>from_model_params(class_count: usize, features: usize) -> Bernoulli { |
| Bernoulli { |
| log_probs: Matrix::zeros(class_count, features), |
| pseudo_count: <span class="number">1f64</span>, |
| } |
| } |
| |
| <span class="kw">fn </span>update_params(<span class="kw-2">&mut </span><span class="self">self</span>, data: <span class="kw-2">&</span>Matrix<f64>, class: usize) -> LearningResult<()> { |
| <span class="kw">let </span>features = data.cols(); |
| |
| <span class="comment">// We add the pseudo count to the class count and feature count |
| </span><span class="kw">let </span>pseudo_cc = data.rows() <span class="kw">as </span>f64 + (<span class="number">2f64 </span>* <span class="self">self</span>.pseudo_count); |
| <span class="kw">let </span>pseudo_fc = data.sum_rows() + <span class="self">self</span>.pseudo_count; |
| |
| <span class="kw">let </span>log_probs = (pseudo_fc.apply(<span class="kw-2">&</span>|x| x.ln()) - pseudo_cc.ln()).into_vec(); |
| |
| <span class="kw">for </span>(i, item) <span class="kw">in </span>log_probs.iter().enumerate().take(features) { |
| <span class="self">self</span>.log_probs[[class, i]] = <span class="kw-2">*</span>item; |
| } |
| |
| <span class="prelude-val">Ok</span>(()) |
| |
| } |
| |
| <span class="kw">fn </span>joint_log_lik(<span class="kw-2">&</span><span class="self">self</span>, |
| data: <span class="kw-2">&</span>Matrix<f64>, |
| class_prior: <span class="kw-2">&</span>[f64]) |
| -> LearningResult<Matrix<f64>> { |
| <span class="kw">let </span>class_count = class_prior.len(); |
| |
| <span class="kw">let </span>neg_prob = <span class="self">self</span>.log_probs.clone().apply(<span class="kw-2">&</span>|x| (<span class="number">1f64 </span>- x.exp()).ln()); |
| |
| <span class="kw">let </span>res = data * (<span class="kw-2">&</span><span class="self">self</span>.log_probs - <span class="kw-2">&</span>neg_prob).transpose(); |
| |
| <span class="comment">// NOTE: Some messy stuff now to get the class row contribution. |
| // Really we want to add to each row the class log-priors and the |
| // neg_prob_sum contribution - the last term in |
| // x log(p) + (1-x)log(1-p) = x (log(p) - log(1-p)) + log(1-p) |
| |
| </span><span class="kw">let </span><span class="kw-2">mut </span>per_class_row = Vec::with_capacity(class_count); |
| <span class="kw">let </span>neg_prob_sum = neg_prob.sum_cols(); |
| |
| <span class="kw">for </span>(idx, p) <span class="kw">in </span>class_prior.into_iter().enumerate() { |
| per_class_row.push(p.ln() + neg_prob_sum[idx]); |
| } |
| |
| <span class="kw">let </span>class_row_mat = Matrix::new(<span class="number">1</span>, class_count, per_class_row); |
| |
| <span class="prelude-val">Ok</span>(res + class_row_mat.select_rows(<span class="kw-2">&</span><span class="macro">vec!</span>[<span class="number">0</span>; data.rows()])) |
| } |
| } |
| |
| <span class="doccomment">/// The Multinomial Naive Bayes model distribution. |
| /// |
| /// Defines: |
| /// |
| /// p(x|C<sub>k</sub>) ∝ ∏<sub>i</sub> p<sub>k</sub><sup>x<sub>i</sub></sup> |
| </span><span class="attribute">#[derive(Debug)] |
| </span><span class="kw">pub struct </span>Multinomial { |
| log_probs: Matrix<f64>, |
| pseudo_count: f64, |
| } |
| |
| <span class="kw">impl </span>Multinomial { |
| <span class="doccomment">/// The log probability matrix. |
| /// |
| /// A matrix of class by feature model log-probabilities. |
| </span><span class="kw">pub fn </span>log_probs(<span class="kw-2">&</span><span class="self">self</span>) -> <span class="kw-2">&</span>Matrix<f64> { |
| <span class="kw-2">&</span><span class="self">self</span>.log_probs |
| } |
| } |
| |
| <span class="kw">impl </span>Distribution <span class="kw">for </span>Multinomial { |
| <span class="kw">fn </span>from_model_params(class_count: usize, features: usize) -> Multinomial { |
| Multinomial { |
| log_probs: Matrix::zeros(class_count, features), |
| pseudo_count: <span class="number">1f64</span>, |
| } |
| } |
| |
| <span class="kw">fn </span>update_params(<span class="kw-2">&mut </span><span class="self">self</span>, data: <span class="kw-2">&</span>Matrix<f64>, class: usize) -> LearningResult<()> { |
| <span class="kw">let </span>features = data.cols(); |
| |
| <span class="kw">let </span>pseudo_fc = data.sum_rows() + <span class="self">self</span>.pseudo_count; |
| <span class="kw">let </span>pseudo_cc = pseudo_fc.sum(); |
| |
| <span class="kw">let </span>log_probs = (pseudo_fc.apply(<span class="kw-2">&</span>|x| x.ln()) - pseudo_cc.ln()).into_vec(); |
| |
| <span class="kw">for </span>(i, item) <span class="kw">in </span>log_probs.iter().enumerate().take(features) { |
| <span class="self">self</span>.log_probs[[class, i]] = <span class="kw-2">*</span>item; |
| } |
| |
| <span class="prelude-val">Ok</span>(()) |
| } |
| |
| <span class="kw">fn </span>joint_log_lik(<span class="kw-2">&</span><span class="self">self</span>, |
| data: <span class="kw-2">&</span>Matrix<f64>, |
| class_prior: <span class="kw-2">&</span>[f64]) |
| -> LearningResult<Matrix<f64>> { |
| <span class="kw">let </span>class_count = class_prior.len(); |
| |
| <span class="kw">let </span>res = data * <span class="self">self</span>.log_probs.transpose(); |
| |
| <span class="kw">let </span><span class="kw-2">mut </span>per_class_row = Vec::with_capacity(class_count); |
| <span class="kw">for </span>p <span class="kw">in </span>class_prior { |
| per_class_row.push(p.ln()); |
| } |
| |
| <span class="kw">let </span>class_row_mat = Matrix::new(<span class="number">1</span>, class_count, per_class_row); |
| |
| <span class="prelude-val">Ok</span>(res + class_row_mat.select_rows(<span class="kw-2">&</span><span class="macro">vec!</span>[<span class="number">0</span>; data.rows()])) |
| } |
| } |
| |
| <span class="attribute">#[cfg(test)] |
| </span><span class="kw">mod </span>tests { |
| <span class="kw">use </span><span class="kw">super</span>::NaiveBayes; |
| <span class="kw">use </span><span class="kw">super</span>::Gaussian; |
| <span class="kw">use </span><span class="kw">super</span>::Bernoulli; |
| <span class="kw">use </span><span class="kw">super</span>::Multinomial; |
| |
| <span class="kw">use </span>learning::SupModel; |
| |
| <span class="kw">use </span>linalg::Matrix; |
| |
| <span class="attribute">#[test] |
| </span><span class="kw">fn </span>test_gaussian() { |
| <span class="kw">let </span>inputs = Matrix::new(<span class="number">6</span>, |
| <span class="number">2</span>, |
| <span class="macro">vec!</span>[<span class="number">1.0</span>, <span class="number">1.1</span>, <span class="number">1.1</span>, <span class="number">0.9</span>, <span class="number">2.2</span>, <span class="number">2.3</span>, <span class="number">2.5</span>, <span class="number">2.7</span>, <span class="number">5.2</span>, <span class="number">4.3</span>, <span class="number">6.2</span>, <span class="number">7.3</span>]); |
| |
| <span class="kw">let </span>targets = Matrix::new(<span class="number">6</span>, |
| <span class="number">3</span>, |
| <span class="macro">vec!</span>[<span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, |
| <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>]); |
| |
| <span class="kw">let </span><span class="kw-2">mut </span>model = NaiveBayes::<Gaussian>::new(); |
| model.train(<span class="kw-2">&</span>inputs, <span class="kw-2">&</span>targets).unwrap(); |
| |
| <span class="kw">let </span>outputs = model.predict(<span class="kw-2">&</span>inputs).unwrap(); |
| <span class="macro">assert_eq!</span>(outputs.into_vec(), targets.into_vec()); |
| } |
| |
| <span class="attribute">#[test] |
| </span><span class="kw">fn </span>test_bernoulli() { |
| <span class="kw">let </span>inputs = Matrix::new(<span class="number">4</span>, |
| <span class="number">3</span>, |
| <span class="macro">vec!</span>[<span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">1.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>]); |
| |
| <span class="kw">let </span>targets = Matrix::new(<span class="number">4</span>, <span class="number">2</span>, <span class="macro">vec!</span>[<span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>]); |
| |
| <span class="kw">let </span><span class="kw-2">mut </span>model = NaiveBayes::<Bernoulli>::new(); |
| model.train(<span class="kw-2">&</span>inputs, <span class="kw-2">&</span>targets).unwrap(); |
| |
| <span class="kw">let </span>outputs = model.predict(<span class="kw-2">&</span>inputs).unwrap(); |
| <span class="macro">assert_eq!</span>(outputs.into_vec(), targets.into_vec()); |
| } |
| |
| <span class="attribute">#[test] |
| </span><span class="kw">fn </span>test_multinomial() { |
| <span class="kw">let </span>inputs = Matrix::new(<span class="number">4</span>, |
| <span class="number">3</span>, |
| <span class="macro">vec!</span>[<span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">5.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">11.0</span>, <span class="number">13.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">12.0</span>, <span class="number">3.0</span>, |
| <span class="number">0.0</span>]); |
| |
| <span class="kw">let </span>targets = Matrix::new(<span class="number">4</span>, <span class="number">2</span>, <span class="macro">vec!</span>[<span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>, <span class="number">0.0</span>, <span class="number">1.0</span>]); |
| |
| <span class="kw">let </span><span class="kw-2">mut </span>model = NaiveBayes::<Multinomial>::new(); |
| model.train(<span class="kw-2">&</span>inputs, <span class="kw-2">&</span>targets).unwrap(); |
| |
| <span class="kw">let </span>outputs = model.predict(<span class="kw-2">&</span>inputs).unwrap(); |
| <span class="macro">assert_eq!</span>(outputs.into_vec(), targets.into_vec()); |
| } |
| } |
| </code></pre></div> |
| </section></div></main><div id="rustdoc-vars" data-root-path="../../../" data-current-crate="rusty_machine" data-themes="ayu,dark,light" data-resource-suffix="" data-rustdoc-version="1.66.0-nightly (5c8bff74b 2022-10-21)" ></div></body></html> |