Use new cursors
This commit is contained in:
parent
a37c3a5e77
commit
2261fe39de
3 changed files with 81 additions and 283 deletions
|
|
@ -12,6 +12,7 @@ use tokio::fs;
|
|||
|
||||
use crate::index::SomethingSupportingLeq;
|
||||
use crate::error::{Error, DecodeErrorKind};
|
||||
use crate::cursor::{ReadCursor, WriteCursor, CursorWithStoreHeader};
|
||||
|
||||
use crate::index::Index;
|
||||
|
||||
|
|
@ -159,15 +160,6 @@ impl <T>SomethingSupportingLeq for Store<T>
|
|||
pub const ROWS_FILE_NAME: &'static str = "rows";
|
||||
|
||||
impl <T>Store<T> {
|
||||
// For debugging.
|
||||
// Moves file cursor to the end.
|
||||
pub async fn read_all_bytes(&mut self) -> std::result::Result<Vec<u8>, std::io::Error>{
|
||||
let mut bytes: Vec<u8> = vec![];
|
||||
let mut cursor = self.cursor(AccessMode::Read).await.map_err(|e| e.to_io_or_panic())?;
|
||||
cursor.file.read_to_end(&mut bytes).await?;
|
||||
Ok(bytes)
|
||||
}
|
||||
|
||||
// ===Creation===
|
||||
pub async fn new(table_folder: &str, number_of_columns: usize, primary_column: Column) -> Result<Self> {
|
||||
let path_to_table = Path::new(table_folder);
|
||||
|
|
@ -234,12 +226,26 @@ impl <T>Store<T> {
|
|||
Ok(store)
|
||||
}
|
||||
|
||||
pub async fn cursor(&self, mode: AccessMode) -> Result<Cursor<T>> {
|
||||
Cursor::new(&self, mode).await
|
||||
// ===Cursors===
|
||||
pub async fn read_cursor(&self) -> Result<ReadCursor<T>>
|
||||
where T: Send
|
||||
{
|
||||
ReadCursor::new(self).await
|
||||
}
|
||||
|
||||
pub async fn garbage_collect(&mut self) -> Result<()> {
|
||||
todo!()
|
||||
pub async fn write_cursor(&mut self) -> Result<WriteCursor<T>>
|
||||
where T: Send
|
||||
{
|
||||
WriteCursor::new(self).await
|
||||
}
|
||||
|
||||
// For debugging.
|
||||
pub async fn read_all_bytes(&mut self) -> std::result::Result<Vec<u8>, std::io::Error>
|
||||
where T: Send
|
||||
{
|
||||
let mut cursor = self.read_cursor().await.map_err(|e| e.to_io_or_panic())?;
|
||||
let bytes = cursor.read_all_bytes().await?;
|
||||
Ok(bytes)
|
||||
}
|
||||
}
|
||||
|
||||
|
|
@ -347,251 +353,6 @@ impl <T>EntryDetailed<T> {
|
|||
}
|
||||
|
||||
|
||||
//=================Cursor==================
|
||||
impl <T>Cursor<T> {
|
||||
pub async fn new(store: &Store<T>, mode: AccessMode) -> Result<Self> {
|
||||
let path_to_rows = Path::new(&store.table_folder).join(ROWS_FILE_NAME);
|
||||
let file: File = match mode {
|
||||
AccessMode::Read =>
|
||||
OpenOptions::new()
|
||||
.read(true)
|
||||
.open(path_to_rows)
|
||||
.await?,
|
||||
|
||||
AccessMode::Write =>
|
||||
OpenOptions::new()
|
||||
.read(true)
|
||||
.write(true)
|
||||
.open(path_to_rows)
|
||||
.await?,
|
||||
};
|
||||
|
||||
let mut cursor = Self {
|
||||
header: store.header.clone(),
|
||||
file,
|
||||
data_type: store.data_type,
|
||||
|
||||
eof_file_position: 0, // This will be overwriten by the seek_to_start_of_data
|
||||
};
|
||||
cursor.seek_to_start_of_data().await?;
|
||||
|
||||
Ok(cursor)
|
||||
}
|
||||
|
||||
//===primitive file operations===
|
||||
// Moves the file cursor right.
|
||||
async fn write_bytes(&mut self, bytes: &[u8]) -> Result<usize> {
|
||||
Ok(self.file.write(bytes).await?)
|
||||
}
|
||||
|
||||
// Moves the file cursor right.
|
||||
async fn read_bytes(&mut self, bytes: &mut [u8]) -> Result<()> {
|
||||
self.file.read_exact(bytes).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
// Moves the file cursor right.
|
||||
async fn get_bytes(&mut self, count: usize) -> Result<Vec<u8>> {
|
||||
let mut result: Vec<u8> = Vec::with_capacity(count);
|
||||
self.read_bytes(&mut result).await?;
|
||||
Ok(result)
|
||||
}
|
||||
|
||||
pub async fn seek_to(&mut self, file_position: FilePosition) -> Result<()> {
|
||||
self.file.seek(SeekFrom::Start(file_position)).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn seek_to_start(&mut self) -> Result<()> {
|
||||
self.file.seek(SeekFrom::Start(0)).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn seek_to_end(&mut self) -> Result<()> {
|
||||
self.file.seek(SeekFrom::End(0)).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn seek_to_start_of_data(&mut self) -> Result<()> {
|
||||
self.seek_to(StoreHeader::SIZE as u64).await
|
||||
}
|
||||
|
||||
pub async fn current_file_position(&mut self) -> Result<FilePosition> {
|
||||
let next_file_position: FilePosition = self.file.stream_position().await?;
|
||||
Ok(next_file_position)
|
||||
}
|
||||
|
||||
async fn is_at_eof(&mut self) -> Result<bool> {
|
||||
Ok(self.current_file_position().await? == self.eof_file_position)
|
||||
}
|
||||
|
||||
pub async fn less_than_eq(&mut self, file_position0: FilePosition, file_position1: FilePosition) -> Result<bool> {
|
||||
todo!()
|
||||
}
|
||||
|
||||
// ===Iteration===
|
||||
// Assumes that the current file position is at a valid entry or EOF.
|
||||
pub async fn next(&mut self) -> Result<Option<EntryDetailed<T>>>
|
||||
where T: Decode
|
||||
{
|
||||
if self.is_at_eof().await? {
|
||||
return Ok(None)
|
||||
}
|
||||
|
||||
let header = self.read_entry_header().await?;
|
||||
|
||||
let mut data_bytes: Vec<u8> = vec![0; header.size_of_data()];
|
||||
self.read_bytes(&mut data_bytes).await?;
|
||||
let entry: EntryDetailed<T> =
|
||||
EntryDetailed::decode(header, self.header.number_of_columns, &mut data_bytes)?;
|
||||
|
||||
Ok(Some(entry))
|
||||
}
|
||||
|
||||
|
||||
// ===Store Header Manipulation===
|
||||
async fn increment_total_count(&mut self) -> Result<()> {
|
||||
self.seek_to_start().await?;
|
||||
self.seek_to(StoreHeader::TOTAL_COUNT_OFFSET as u64).await?;
|
||||
let new_count = self.header.increment_total_count();
|
||||
self.write_bytes(&encode::<usize>(&new_count)?).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
async fn increment_deleted_count(&mut self) -> Result<()> {
|
||||
self.seek_to_start().await?;
|
||||
self.seek_to(StoreHeader::DELETED_COUNT_OFFSET as u64).await?;
|
||||
let new_count = self.header.increment_deleted_count();
|
||||
self.write_bytes(&encode::<usize>(&new_count)?).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
// ===Entry Header Manipulation===
|
||||
// assumes we are at the start of the valid entry.
|
||||
async fn set_entry_is_deleted_to(&mut self, is_deleted: bool) -> Result<()> {
|
||||
self.seek_to(EntryHeaderWithDataSize::IS_DELETED_OFFSET as u64).await?;
|
||||
self.write_bytes(&encode::<bool>(&is_deleted)?).await?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
// ===Append Entry===
|
||||
|
||||
// Moves cursor to the end.
|
||||
// Returns file position to the start of the new entry.
|
||||
pub async fn append_entry(&mut self, entry: &Entry<T>) -> Result<FilePosition>
|
||||
where T: Encode
|
||||
{
|
||||
self.increment_total_count().await?;
|
||||
|
||||
let encoded_entry: Vec<u8> = entry.encode()?;
|
||||
self.seek_to_end().await?;
|
||||
let file_position: FilePosition = self.current_file_position().await?;
|
||||
self.write_bytes(&encoded_entry).await?;
|
||||
|
||||
let eof_file_position: FilePosition = self.current_file_position().await?;
|
||||
self.eof_file_position = eof_file_position;
|
||||
|
||||
Ok(file_position)
|
||||
}
|
||||
|
||||
// ===Deletion===
|
||||
pub async fn mark_deleted_at(&mut self, file_position: FilePosition) -> Result<()> {
|
||||
self.seek_to(file_position).await?;
|
||||
let entry_header = self.read_entry_header().await?;
|
||||
if entry_header.is_deleted {
|
||||
Ok(())
|
||||
} else {
|
||||
self.increment_deleted_count().await?;
|
||||
self.seek_to(file_position).await?;
|
||||
self.set_entry_is_deleted_to(true).await?;
|
||||
|
||||
self.attempt_garbage_collection_if_necessary().await?;
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
async fn attempt_garbage_collection_if_necessary(&mut self) -> Result<()> {
|
||||
// TODO: What should be the policy? Counting size of garbage? Counting how many entries are
|
||||
// garbage?
|
||||
if self.header.deleted_count > 100 {
|
||||
todo!()
|
||||
} else {
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
// ===Lookup===
|
||||
// WARNING: The cursor has to be at the start of an entry. Otherwise garbage data will be
|
||||
// decoded as an entry.
|
||||
async fn read_entry_header(&mut self) -> Result<EntryHeaderWithDataSize> {
|
||||
let number_of_columns: usize = self.header.number_of_columns;
|
||||
let mut header_bytes: Vec<u8> = vec![0; EntryHeaderWithDataSize::size(number_of_columns)];
|
||||
self.read_bytes(&mut header_bytes).await?;
|
||||
let header = EntryHeaderWithDataSize::decode(&mut header_bytes[..], number_of_columns)?;
|
||||
// TODO: Get rid of the println's
|
||||
// println!("HEADER_BYTES: {:?}", header_bytes);
|
||||
// println!("HEADER: {:?}", header);
|
||||
|
||||
Ok(header)
|
||||
}
|
||||
|
||||
pub async fn read_entry_header_at(&mut self, file_position: FilePosition) -> Result<EntryHeaderWithDataSize> {
|
||||
self.seek_to(file_position).await?;
|
||||
self.read_entry_header().await
|
||||
}
|
||||
|
||||
pub async fn search_for(&mut self, index: T) -> Result<()>
|
||||
where T: Send
|
||||
{
|
||||
// let index = self.primary_index.borrow_mut();
|
||||
// let x = index.lookup(self, 123).await?;
|
||||
todo!()
|
||||
}
|
||||
|
||||
|
||||
// Returns None when file_positoin == eof_file_position
|
||||
pub async fn read_entry_at(&mut self, file_position: FilePosition) -> Result<Option<EntryDetailed<T>>>
|
||||
where T: Decode
|
||||
{
|
||||
self.seek_to(file_position).await?;
|
||||
self.next().await
|
||||
}
|
||||
|
||||
// TODO: This needs to be some sort of an iterator
|
||||
// pub async fn entries() -> EntryIterator<T> {
|
||||
// todo!()
|
||||
// }
|
||||
|
||||
pub async fn read_entries(&mut self) -> Result<()>
|
||||
where T: Decode + std::fmt::Debug
|
||||
{
|
||||
self.seek_to_start_of_data().await?;
|
||||
let mut file_position: FilePosition = self.current_file_position().await?;
|
||||
loop {
|
||||
match self.read_entry_at(file_position).await? {
|
||||
Some(entry) => {
|
||||
println!("{:?}", entry);
|
||||
file_position = self.current_file_position().await?;
|
||||
},
|
||||
None => {
|
||||
println!("END of entries.");
|
||||
return Ok(())
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn search_for_entry_with_id(&mut self, id: T) -> Result<Option<EntryDetailed<T>>> {
|
||||
// TODO: make call to the primary index
|
||||
todo!()
|
||||
}
|
||||
|
||||
// TODO: This needs to be some sort of an iterator
|
||||
pub async fn get_all_eq(&self, column: Column, value: T) -> Result<Option<EntryDetailed<T>>> {
|
||||
todo!()
|
||||
}
|
||||
}
|
||||
|
||||
// impl StorageEngine for ColumnStore {
|
||||
// async fn append(&mut self, id: Index, entry: Row<T>) -> Result<???, Error>
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue