| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| package org.apache.hyracks.storage.am.bloomfilter.impls; |
| |
| import java.nio.ByteBuffer; |
| |
| import org.apache.hyracks.api.exceptions.HyracksDataException; |
| import org.apache.hyracks.api.io.FileReference; |
| import org.apache.hyracks.dataflow.common.data.accessors.ITupleReference; |
| import org.apache.hyracks.storage.am.common.api.IIndexBulkLoader; |
| import org.apache.hyracks.storage.am.common.api.IndexException; |
| import org.apache.hyracks.storage.common.buffercache.IBufferCache; |
| import org.apache.hyracks.storage.common.buffercache.ICachedPage; |
| import org.apache.hyracks.storage.common.file.BufferedFileHandle; |
| import org.apache.hyracks.storage.common.file.IFileMapProvider; |
| |
| public class BloomFilter { |
| |
| private final static int METADATA_PAGE_ID = 0; |
| private final static int NUM_PAGES_OFFSET = 0; // 0 |
| private final static int NUM_HASHES_USED_OFFSET = NUM_PAGES_OFFSET + 4; // 4 |
| private final static int NUM_ELEMENTS_OFFSET = NUM_HASHES_USED_OFFSET + 4; // 8 |
| private final static int NUM_BITS_OFFSET = NUM_ELEMENTS_OFFSET + 8; // 12 |
| |
| private final IBufferCache bufferCache; |
| private final IFileMapProvider fileMapProvider; |
| private final FileReference file; |
| private final int[] keyFields; |
| private int fileId = -1; |
| private boolean isActivated = false; |
| |
| private int numPages; |
| private int numHashes; |
| private long numElements; |
| private long numBits; |
| private final int numBitsPerPage; |
| private final static byte[] ZERO_BUFFER = new byte[131072]; // 128kb |
| private final static long SEED = 0L; |
| |
| public BloomFilter(IBufferCache bufferCache, IFileMapProvider fileMapProvider, FileReference file, int[] keyFields) |
| throws HyracksDataException { |
| this.bufferCache = bufferCache; |
| this.fileMapProvider = fileMapProvider; |
| this.file = file; |
| this.keyFields = keyFields; |
| this.numBitsPerPage = bufferCache.getPageSize() * Byte.SIZE; |
| } |
| |
| public int getFileId() { |
| return fileId; |
| } |
| |
| public FileReference getFileReference() { |
| return file; |
| } |
| |
| public int getNumPages() throws HyracksDataException { |
| if (!isActivated) { |
| throw new HyracksDataException("The bloom filter is not activated."); |
| } |
| return numPages; |
| } |
| |
| public long getNumElements() throws HyracksDataException { |
| if (!isActivated) { |
| throw new HyracksDataException("The bloom filter is not activated."); |
| } |
| return numElements; |
| } |
| |
| public boolean contains(ITupleReference tuple, long[] hashes) throws HyracksDataException { |
| if (numPages == 0) { |
| return false; |
| } |
| MurmurHash128Bit.hash3_x64_128(tuple, keyFields, SEED, hashes); |
| for (int i = 0; i < numHashes; ++i) { |
| long hash = Math.abs((hashes[0] + i * hashes[1]) % numBits); |
| |
| // we increment the page id by one, since the metadata page id of the filter is 0. |
| ICachedPage page = bufferCache.pin( |
| BufferedFileHandle.getDiskPageId(fileId, (int) (hash / numBitsPerPage) + 1), false); |
| page.acquireReadLatch(); |
| try { |
| ByteBuffer buffer = page.getBuffer(); |
| int byteIndex = (int) (hash % numBitsPerPage) >> 3; // divide by 8 |
| byte b = buffer.get(byteIndex); |
| int bitIndex = (int) (hash % numBitsPerPage) & 0x07; // mod 8 |
| |
| if (!((b & (1L << bitIndex)) != 0)) { |
| return false; |
| } |
| } finally { |
| page.releaseReadLatch(); |
| bufferCache.unpin(page); |
| } |
| |
| } |
| return true; |
| } |
| |
| private void prepareFile() throws HyracksDataException { |
| boolean fileIsMapped = false; |
| synchronized (fileMapProvider) { |
| fileIsMapped = fileMapProvider.isMapped(file); |
| if (!fileIsMapped) { |
| bufferCache.createFile(file); |
| } |
| fileId = fileMapProvider.lookupFileId(file); |
| try { |
| // Also creates the file if it doesn't exist yet. |
| bufferCache.openFile(fileId); |
| } catch (HyracksDataException e) { |
| // Revert state of buffer cache since file failed to open. |
| if (!fileIsMapped) { |
| bufferCache.deleteFile(fileId, false); |
| } |
| throw e; |
| } |
| } |
| } |
| |
| public synchronized void create() throws HyracksDataException { |
| if (isActivated) { |
| throw new HyracksDataException("Failed to create the bloom filter since it is activated."); |
| } |
| prepareFile(); |
| ICachedPage metaPage = bufferCache.pin(BufferedFileHandle.getDiskPageId(fileId, METADATA_PAGE_ID), true); |
| metaPage.acquireWriteLatch(); |
| try { |
| metaPage.getBuffer().putInt(NUM_PAGES_OFFSET, 0); |
| metaPage.getBuffer().putInt(NUM_HASHES_USED_OFFSET, 0); |
| metaPage.getBuffer().putLong(NUM_ELEMENTS_OFFSET, 0L); |
| metaPage.getBuffer().putLong(NUM_BITS_OFFSET, 0L); |
| } finally { |
| metaPage.releaseWriteLatch(true); |
| bufferCache.unpin(metaPage); |
| } |
| bufferCache.closeFile(fileId); |
| } |
| |
| public synchronized void activate() throws HyracksDataException { |
| if (isActivated) { |
| return; |
| } |
| |
| prepareFile(); |
| readBloomFilterMetaData(); |
| isActivated = true; |
| } |
| |
| private void readBloomFilterMetaData() throws HyracksDataException { |
| ICachedPage metaPage = bufferCache.pin(BufferedFileHandle.getDiskPageId(fileId, METADATA_PAGE_ID), false); |
| metaPage.acquireReadLatch(); |
| try { |
| numPages = metaPage.getBuffer().getInt(NUM_PAGES_OFFSET); |
| numHashes = metaPage.getBuffer().getInt(NUM_HASHES_USED_OFFSET); |
| numElements = metaPage.getBuffer().getLong(NUM_ELEMENTS_OFFSET); |
| numBits = metaPage.getBuffer().getLong(NUM_BITS_OFFSET); |
| } finally { |
| metaPage.releaseReadLatch(); |
| bufferCache.unpin(metaPage); |
| } |
| } |
| |
| public synchronized void deactivate() throws HyracksDataException { |
| if (!isActivated) { |
| return; |
| } |
| bufferCache.closeFile(fileId); |
| isActivated = false; |
| } |
| |
| public synchronized void destroy() throws HyracksDataException { |
| if (isActivated) { |
| throw new HyracksDataException("Failed to destroy the bloom filter since it is activated."); |
| } |
| |
| file.delete(); |
| if (fileId == -1) { |
| return; |
| } |
| bufferCache.deleteFile(fileId, false); |
| fileId = -1; |
| } |
| |
| public IIndexBulkLoader createBuilder(long numElements, int numHashes, int numBitsPerElement) |
| throws HyracksDataException { |
| return new BloomFilterBuilder(numElements, numHashes, numBitsPerElement); |
| } |
| |
| public class BloomFilterBuilder implements IIndexBulkLoader { |
| private final long[] hashes = new long[2]; |
| private final long numElements; |
| private final int numHashes; |
| private final long numBits; |
| private final int numPages; |
| |
| public BloomFilterBuilder(long numElements, int numHashes, int numBitsPerElement) throws HyracksDataException { |
| if (!isActivated) { |
| throw new HyracksDataException("Failed to create the bloom filter builder since it is not activated."); |
| } |
| |
| this.numElements = numElements; |
| this.numHashes = numHashes; |
| numBits = this.numElements * numBitsPerElement; |
| long tmp = (long) Math.ceil(numBits / (double) numBitsPerPage); |
| if (tmp > Integer.MAX_VALUE) { |
| throw new HyracksDataException("Cannot create a bloom filter with his huge number of pages."); |
| } |
| numPages = (int) tmp; |
| persistBloomFilterMetaData(); |
| readBloomFilterMetaData(); |
| int currentPageId = 1; |
| while (currentPageId <= numPages) { |
| ICachedPage page = bufferCache.pin(BufferedFileHandle.getDiskPageId(fileId, currentPageId), true); |
| page.acquireWriteLatch(); |
| try { |
| initPage(page.getBuffer().array()); |
| } finally { |
| page.releaseWriteLatch(true); |
| bufferCache.unpin(page); |
| } |
| ++currentPageId; |
| } |
| } |
| |
| private void initPage(byte[] array) { |
| int numRounds = array.length / ZERO_BUFFER.length; |
| int leftOver = array.length % ZERO_BUFFER.length; |
| int destPos = 0; |
| for (int i = 0; i < numRounds; i++) { |
| System.arraycopy(ZERO_BUFFER, 0, array, destPos, ZERO_BUFFER.length); |
| destPos = (i + 1) * ZERO_BUFFER.length; |
| } |
| if (leftOver > 0) { |
| System.arraycopy(ZERO_BUFFER, 0, array, destPos, leftOver); |
| } |
| } |
| |
| private void persistBloomFilterMetaData() throws HyracksDataException { |
| ICachedPage metaPage = bufferCache.pin(BufferedFileHandle.getDiskPageId(fileId, METADATA_PAGE_ID), false); |
| metaPage.acquireWriteLatch(); |
| try { |
| metaPage.getBuffer().putInt(NUM_PAGES_OFFSET, numPages); |
| metaPage.getBuffer().putInt(NUM_HASHES_USED_OFFSET, numHashes); |
| metaPage.getBuffer().putLong(NUM_ELEMENTS_OFFSET, numElements); |
| metaPage.getBuffer().putLong(NUM_BITS_OFFSET, numBits); |
| } finally { |
| metaPage.releaseWriteLatch(true); |
| bufferCache.unpin(metaPage); |
| } |
| } |
| |
| @Override |
| public void add(ITupleReference tuple) throws IndexException, HyracksDataException { |
| if (numPages == 0) { |
| throw new HyracksDataException( |
| "Cannot add elements to this filter since it is supposed to be empty (number of elements hint passed to the filter during construction was 0)."); |
| } |
| MurmurHash128Bit.hash3_x64_128(tuple, keyFields, SEED, hashes); |
| for (int i = 0; i < numHashes; ++i) { |
| long hash = Math.abs((hashes[0] + i * hashes[1]) % numBits); |
| |
| // we increment the page id by one, since the metadata page id of the filter is 0. |
| ICachedPage page = bufferCache.pin( |
| BufferedFileHandle.getDiskPageId(fileId, (int) (hash / numBitsPerPage) + 1), false); |
| page.acquireWriteLatch(); |
| try { |
| ByteBuffer buffer = page.getBuffer(); |
| int byteIndex = (int) (hash % numBitsPerPage) >> 3; // divide by 8 |
| byte b = buffer.array()[byteIndex]; |
| int bitIndex = (int) (hash % numBitsPerPage) & 0x07; // mod 8 |
| b = (byte) (b | (1 << bitIndex)); |
| buffer.array()[byteIndex] = b; |
| } finally { |
| page.releaseWriteLatch(true); |
| bufferCache.unpin(page); |
| } |
| } |
| } |
| |
| @Override |
| public void end() throws HyracksDataException, IndexException { |
| } |
| |
| } |
| } |