mirror of https://github.com/voidlizard/hbs2
293 lines
7.6 KiB
Haskell
293 lines
7.6 KiB
Haskell
module HBS2.Actors.ChunkWriter
|
|
( ChunkWriter
|
|
, ChunkId
|
|
, newChunkWriterIO
|
|
, runChunkWriter
|
|
, stopChunkWriter
|
|
, delBlock
|
|
, commitBlock
|
|
, writeChunk
|
|
, getHash
|
|
, blocksInProcess
|
|
) where
|
|
|
|
import HBS2.Prelude
|
|
import HBS2.Actors
|
|
import HBS2.Hash
|
|
import HBS2.Storage
|
|
import HBS2.Defaults
|
|
import HBS2.Clock
|
|
|
|
import Data.List qualified as L
|
|
import Data.Functor
|
|
import Data.Function
|
|
import Control.Exception
|
|
import Data.ByteString.Lazy (ByteString)
|
|
import Data.ByteString.Lazy qualified as B
|
|
-- import Data.Cache (Cache)
|
|
-- import Data.Cache qualified as Cache
|
|
import Data.Foldable
|
|
import Data.Traversable
|
|
import Data.Hashable (hash)
|
|
import Data.Maybe
|
|
import Data.Word
|
|
import Prettyprinter
|
|
import System.Directory
|
|
import System.FilePath
|
|
import System.IO.Error
|
|
import System.IO
|
|
import System.IO.Temp
|
|
import System.FileLock
|
|
|
|
import Control.Concurrent.Async
|
|
|
|
import Data.Cache (Cache)
|
|
import Data.Cache qualified as Cache
|
|
import Control.Concurrent.STM
|
|
import Control.Concurrent.STM.TVar as TV
|
|
import Control.Concurrent.STM.TBQueue qualified as Q
|
|
import Control.Concurrent.STM.TSem qualified as Sem
|
|
import Control.Concurrent.STM.TSem (TSem)
|
|
|
|
import Control.Concurrent.STM.TQueue qualified as Q0
|
|
import Control.Concurrent
|
|
|
|
import Data.HashMap.Strict (HashMap)
|
|
import Data.HashMap.Strict qualified as HashMap
|
|
|
|
--
|
|
--
|
|
--TODO: cache file handles
|
|
|
|
newtype ChunkId = ChunkId FilePath
|
|
deriving newtype (IsString)
|
|
deriving stock (Eq,Ord,Show)
|
|
|
|
data ChunkWriter h m = forall a . ( MonadIO m
|
|
, Storage a h ByteString m
|
|
, Block ByteString ~ ByteString
|
|
) =>
|
|
ChunkWriter
|
|
{ stopped :: TVar Bool
|
|
, pipeline :: Pipeline m ()
|
|
, dir :: FilePath
|
|
, storage :: a
|
|
, perBlock :: TVar (HashMap FilePath [IO ()])
|
|
, perBlockSem :: TVar (HashMap FilePath TSem)
|
|
}
|
|
|
|
|
|
blocksInProcess :: MonadIO m => ChunkWriter h m -> m Int
|
|
blocksInProcess cw = undefined
|
|
-- liftIO $ Cache.purgeExpired cache >> Cache.size cache
|
|
-- where
|
|
-- cache = perBlock cw
|
|
|
|
runChunkWriter :: forall h m . ( Eq (Hash h)
|
|
, Hashable (Hash h)
|
|
, MonadIO m )
|
|
=> ChunkWriter h m -> m ()
|
|
|
|
runChunkWriter = runChunkWriter2
|
|
|
|
|
|
runChunkWriter2 :: forall h m . ( Eq (Hash h)
|
|
, Hashable (Hash h)
|
|
, MonadIO m )
|
|
=> ChunkWriter h m -> m ()
|
|
|
|
runChunkWriter2 w = do
|
|
liftIO $ createDirectoryIfMissing True ( dir w )
|
|
let tv = perBlock w
|
|
fix \next -> do
|
|
keys <- liftIO $ readTVarIO tv <&> (L.take 10 . HashMap.keys)
|
|
for_ keys (flush w)
|
|
pause ( 0.25 :: Timeout 'Seconds)
|
|
next
|
|
|
|
stopChunkWriter :: MonadIO m => ChunkWriter h m -> m ()
|
|
stopChunkWriter w = do
|
|
liftIO $ atomically $ writeTVar (stopped w) True
|
|
|
|
newChunkWriterIO :: forall h a m . ( Key h ~ Hash h, h ~ HbSync
|
|
, Storage a h ByteString m
|
|
, Block ByteString ~ ByteString
|
|
, MonadIO m
|
|
)
|
|
=> a
|
|
-> Maybe FilePath
|
|
-> m (ChunkWriter h m)
|
|
|
|
newChunkWriterIO s tmp = do
|
|
pip <- newPipeline defChunkWriterQ
|
|
|
|
def <- liftIO $ getXdgDirectory XdgData (defStorePath </> "temp-chunks")
|
|
let d = fromMaybe def tmp
|
|
|
|
mt <- liftIO $ newTVarIO mempty
|
|
mts <- liftIO $ newTVarIO mempty
|
|
|
|
running <- liftIO $ newTVarIO False
|
|
|
|
pure $
|
|
ChunkWriter
|
|
{ stopped = running
|
|
, pipeline = pip
|
|
, dir = d
|
|
, storage = s
|
|
, perBlock = mt
|
|
, perBlockSem = mts
|
|
}
|
|
|
|
makeFileName :: (Hashable salt, Pretty (Hash h)) => ChunkWriter h m -> salt -> Hash h -> FilePath
|
|
makeFileName w salt h = dir w </> suff
|
|
where
|
|
suff = show $ pretty (fromIntegral (hash salt) :: Word32) <> "@" <> pretty h
|
|
|
|
delBlock :: (Hashable salt, MonadIO m, Pretty (Hash h))
|
|
=> ChunkWriter h m -> salt -> Hash h -> m ()
|
|
|
|
delBlock w salt h = liftIO do
|
|
void $ tryJust (guard . isDoesNotExistError) (removeFile fn)
|
|
where
|
|
fn = makeFileName w salt h
|
|
|
|
writeChunk :: ( Hashable salt
|
|
, MonadIO m
|
|
, Pretty (Hash h)
|
|
, Hashable (Hash h), Eq (Hash h)
|
|
)
|
|
=> ChunkWriter h m
|
|
-> salt
|
|
-> Hash h
|
|
-> Offset
|
|
-> ByteString -> m ()
|
|
|
|
writeChunk = writeChunk2
|
|
|
|
|
|
getHash :: forall salt h m .
|
|
( Hashable salt
|
|
, Hashed h ByteString
|
|
, MonadIO m
|
|
, Block ByteString ~ ByteString
|
|
, Pretty (Hash h)
|
|
, Hashable (Hash h), Eq (Hash h)
|
|
)
|
|
=> ChunkWriter h m
|
|
-> salt
|
|
-> Hash h
|
|
-> m (Hash h)
|
|
|
|
getHash = getHash2
|
|
|
|
|
|
commitBlock :: forall salt h m .
|
|
( Hashable salt
|
|
, Hashed h ByteString
|
|
, Block ByteString ~ ByteString
|
|
, MonadIO m
|
|
, Pretty (Hash h)
|
|
, Hashable (Hash h), Eq (Hash h)
|
|
)
|
|
=> ChunkWriter h m
|
|
-> salt
|
|
-> Hash h
|
|
-> m ()
|
|
|
|
commitBlock = commitBlock2
|
|
|
|
writeChunk2 :: (Hashable salt, MonadIO m, Pretty (Hash h), Hashable (Hash h), Eq (Hash h))
|
|
=> ChunkWriter h m
|
|
-> salt
|
|
-> Hash h
|
|
-> Offset
|
|
-> ByteString -> m ()
|
|
|
|
writeChunk2 w salt h o bs = do
|
|
|
|
let cache = perBlock w
|
|
|
|
let action = do
|
|
withBinaryFile fn ReadWriteMode $ \fh -> do
|
|
hSeek fh AbsoluteSeek (fromIntegral o)
|
|
B.hPutStr fh bs
|
|
|
|
liftIO $ do
|
|
atomically $ modifyTVar cache (HashMap.insertWith (<>) fn [action])
|
|
|
|
where
|
|
fn = makeFileName w salt h
|
|
|
|
flush w fn = do
|
|
let cache = perBlock w
|
|
let sems = perBlockSem w
|
|
|
|
liftIO $ do
|
|
|
|
nsem <- atomically $ Sem.newTSem 1
|
|
sem <- atomically $ stateTVar sems $ \hm -> let found = HashMap.lookup fn hm
|
|
in case found of
|
|
Nothing -> (nsem, HashMap.insert fn nsem hm)
|
|
Just s -> (s, hm)
|
|
atomically $ Sem.waitTSem sem
|
|
|
|
actions <- atomically $ stateTVar cache (\v -> (HashMap.lookup fn v, HashMap.delete fn v))
|
|
sequence_ (fromMaybe mempty actions)
|
|
|
|
atomically $ Sem.signalTSem sem
|
|
|
|
-- Blocking!
|
|
-- we need to write last chunk before this will happen
|
|
-- FIXME: incremental calculation,
|
|
-- streaming, blah-blah
|
|
getHash2 :: forall salt h m .
|
|
( Hashable salt
|
|
, Hashed h ByteString
|
|
, MonadIO m
|
|
, Block ByteString ~ ByteString
|
|
, Pretty (Hash h)
|
|
, Hashable (Hash h), Eq (Hash h)
|
|
)
|
|
=> ChunkWriter h m
|
|
-> salt
|
|
-> Hash h
|
|
-> m (Hash h)
|
|
|
|
getHash2 w salt h = do
|
|
flush w fn
|
|
liftIO $ hashObject @h <$> B.readFile fn
|
|
|
|
where
|
|
fn = makeFileName w salt h
|
|
|
|
|
|
commitBlock2 :: forall salt h m .
|
|
( Hashable salt
|
|
, Hashed h ByteString
|
|
, Block ByteString ~ ByteString
|
|
, MonadIO m
|
|
, Pretty (Hash h)
|
|
, Hashable (Hash h), Eq (Hash h)
|
|
)
|
|
=> ChunkWriter h m
|
|
-> salt
|
|
-> Hash h
|
|
-> m ()
|
|
|
|
commitBlock2 w@(ChunkWriter {storage = stor}) salt h = do
|
|
let cache = perBlock w
|
|
let se = perBlockSem w
|
|
flush w fn
|
|
s <- liftIO $ B.readFile fn
|
|
void $ putBlock stor s
|
|
delBlock w salt h
|
|
|
|
liftIO $ atomically $ TV.modifyTVar' cache $ HashMap.delete fn
|
|
liftIO $ atomically $ TV.modifyTVar' se $ HashMap.delete fn
|
|
|
|
where
|
|
fn = makeFileName w salt h
|
|
|
|
|