code
stringlengths 5
1.03M
| repo_name
stringlengths 5
90
| path
stringlengths 4
158
| license
stringclasses 15
values | size
int64 5
1.03M
| n_ast_errors
int64 0
53.9k
| ast_max_depth
int64 2
4.17k
| n_whitespaces
int64 0
365k
| n_ast_nodes
int64 3
317k
| n_ast_terminals
int64 1
171k
| n_ast_nonterminals
int64 1
146k
| loc
int64 -1
37.3k
| cycloplexity
int64 -1
1.31k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE PatternGuards #-}
module Main (main) where
import Codec.Compression.Zlib (compress)
import Control.Applicative ((<$>))
import Control.Concurrent (forkIO)
import Control.Concurrent.MVar
import qualified Control.Exception as C
import Control.Monad (forM, when)
import qualified Data.Attoparsec.ByteString.Char8 as A
import Data.Attoparsec.ByteString.Lazy (parse, Result(..))
import Data.Binary.Put
import Data.Bits
import Data.ByteString (ByteString)
import qualified Data.ByteString as B
import qualified Data.ByteString.Base16 as B16
import qualified Data.ByteString.Char8 as BC
import qualified Data.ByteString.Lazy as L
import qualified Data.ByteString.Lazy.Char8 as LC
import qualified Data.ByteString.Lazy.Internal as I
import Data.Digest.Pure.SHA (bytestringDigest, sha1)
import Data.Int
import Data.IORef
import Data.List (groupBy, nubBy, sortBy)
import Data.Thyme.Clock (UTCTime)
import Data.Thyme.Format (formatTime)
import Data.Thyme.Time () -- For instance Num POSIXTime (a.k.a. NominalDiffTime)
import Data.Thyme.Time.Core (posixSecondsToUTCTime)
import Data.Word
import System.Directory (canonicalizePath, createDirectoryIfMissing, doesDirectoryExist)
import System.Environment (getArgs)
import System.Exit (ExitCode(..))
import System.FilePath (joinPath, splitDirectories, (</>))
import System.IO
( hClose, hFlush, hGetContents, hPutStr, hSeek, hSetFileSize, openFile
, Handle, IOMode(..), SeekMode(AbsoluteSeek)
)
import qualified System.IO.Streams as S
import System.IO.Unsafe (unsafeInterleaveIO)
import System.Locale (defaultTimeLocale)
import System.Process
( createProcess, proc, readProcessWithExitCode, waitForProcess
, CreateProcess(..), StdStream(..)
)
----------------------------------------------------------------------
-- Command-line
----------------------------------------------------------------------
main :: IO ()
main = do
args <- getArgs
case args of
["init", gitDir_] -> do
gitDir <- canonicalizePath gitDir_
e <- doesDirectoryExist gitDir
if e
then error $ "directory " ++ gitDir ++ " already exist"
else initRepository gitDir
["ensure", gitDir_] -> do
gitDir <- canonicalizePath gitDir_
ensureGitDir gitDir
["test"] -> do
gitDir <- canonicalizePath ".git"
readRefs gitDir Nothing >>= print
readHeads gitDir Nothing >>= print
readObjects gitDir
[ Ref "32ab47487f560b11fdc758eedd9b43ee7aeb8684" -- blob
, Ref "740fc0e4923e9f1ee5e0488cb1e7877c990a3f69" -- tree
, Ref "a5b6d23259c76b66c933ba9940f6afcdf1bf3fff" -- commit
] >>= print
revList gitDir (Ref "a5b6d23259c76b66c933ba9940f6afcdf1bf3fff") >>= print
readRevs gitDir (Ref "a5b6d23259c76b66c933ba9940f6afcdf1bf3fff") >>= print
["ls", gitDir_] -> do
gitDir <- canonicalizePath gitDir_
putStrLn "blob" -- access logical blobs, not chunks
putStrLn "branch"
putStrLn "chunk" -- access all blobs
putStrLn "commit"
putStrLn "time" -- same as commit but using the bup format YY-MM-DD-hhmmss instead of a hash
["ls", gitDir_, path] -> do
gitDir <- canonicalizePath gitDir_
case splitDirectories path of
["branch"] -> do
hds <- readHeads gitDir Nothing
mapM_ (BC.putStrLn . fst) hds
["branch", branch] -> do
enterBranch gitDir branch
"branch" : branch : commit : rest -> enterBranchCommit gitDir ls branch commit rest
["cat", gitDir_, path] -> do
gitDir <- canonicalizePath gitDir_
case splitDirectories path of
"branch" : branch : commit : rest -> enterBranchCommit gitDir cat branch commit rest
["pack", gitDir_] -> do
gitDir <- canonicalizePath gitDir_
(sha, tree) <- repack gitDir "objects/pack/p1.pack"
[L "README.md" $ Blob 15 "Nice isn't it?\n"]
Nothing Nothing
"Initial commit."
"refs/heads/master"
(sha2, tree2) <- repack gitDir "objects/pack/p2.pack"
[L "new.txt" $ Blob 6 "hello\n"]
(Just sha) (Just tree)
"Added new.txt."
"refs/heads/new-branch"
(sha3, tree3) <- repack gitDir "objects/pack/p3.pack"
[T "a" [T "b" [L "c.txt" $ Blob 7 "Super.\n"]]]
(Just sha2) (Just tree2)
"Added a/b/c.txt."
"refs/heads/branch-3"
_ <- repack gitDir "objects/pack/p4.pack"
[L "new.txt" $ Blob 4 "bye\n"]
(Just sha3) (Just tree3)
"Changed hello to bye."
"refs/heads/branch-4"
let blob bs = Blob (L.length bs) bs
_ <- repack gitDir "objects/pack/p5.pack"
(groupBlobs
[ ("README.md", blob "Pack 5\n")
, ("bin/script.hs", blob "main = putStrLn \"Hello, world!\"\n")
, ("tests/data/set-1/file-00.txt", blob "10\n")
, ("tests/data/set-1/file-01.txt", blob "11\n")
, ("tests/data/EMPTY", blob "")
, ("tests/data/set-2/file-00.txt", blob "20\n")
, ("tests/data/set-1/file-02.txt", blob "12\n")
])
Nothing Nothing
"Initial commit."
"refs/heads/branch-5"
return ()
-- `buh fast-export` is meant to be used with `buh fast-import --files` i.e.
-- the commit command is not issued.
-- Example:
-- buh fast-export .git branch/develop/latest/static/css | buh fast-import --files barerepo
-- This will create a css directory at the root of the bare repo.
["fast-export", gitDir_, path] -> do
gitDir <- canonicalizePath gitDir_
checkRepository gitDir
case splitDirectories path of
"branch" : branch : commit : rest -> enterBranchCommit gitDir export branch commit rest
["fast-import", gitDir_] -> do
ensureGitDir gitDir_
gitDir <- canonicalizePath gitDir_
gitFastImport gitDir False
["fast-import", "--files", gitDir_] -> do
ensureGitDir gitDir_
gitDir <- canonicalizePath gitDir_
gitFastImport gitDir True
xs -> error $ "TODO " ++ show xs
initRepository path = do
(_, _, _, p) <- createProcess (proc "git"
[ "init", "--bare", path
])
_ <- waitForProcess p
return ()
-- | TODO Make the check more comprehensive.
checkRepository path = do
e <- doesDirectoryExist path
e' <- doesDirectoryExist $ path </> "objects"
e'' <- doesDirectoryExist $ path </> "refs"
if e && e' && e''
then return ()
else error "Not a Git repository."
ensureGitDir gitDir = do
e <- doesDirectoryExist gitDir
if e
then return ()
else do
createDirectoryIfMissing True gitDir
initRepository gitDir
-- | If doCommit is True, the commit command is automatically sent before the
-- rest of the input stream.
gitFastImport gitDir doCommit = do
is <- if doCommit
then do
-- If the ref already exists, then continue the commit from it.
refs <- readHeads gitDir $ Just $ Ref "refs/heads/fast-import"
commit <- if length refs == 1
then S.fromLazyByteString . runPut $ feCommit (Just "refs/heads/fast-import")
else S.fromLazyByteString . runPut $ feCommit Nothing
S.appendInputStream commit S.stdin
else return S.stdin
(Just hin, _, _, p) <- createProcess (proc "git"
[ "fast-import", "--date-format=now"
]) { env = Just [("GIT_DIR", gitDir)]
, std_in = CreatePipe
}
sIn <- S.handleToOutputStream hin >>=
S.atEndOfOutput (hClose hin)
S.connect is sIn
_ <- waitForProcess p
return ()
enterBranch :: FilePath -> String -> IO ()
enterBranch gitDir branch = do
hds <- readHeads gitDir Nothing
Sha sha <- lookupPath (BC.pack branch) hds
cs <- readRevs gitDir (Ref sha)
mapM_ (BC.putStrLn . fst) cs
enterBranchCommit :: FilePath -> (FilePath -> String -> Object -> IO a) -> String -> String -> [String] -> IO a
enterBranchCommit gitDir f branch commit path = do
hds <- readHeads gitDir Nothing
Sha sha <- lookupPath (BC.pack branch) hds
cs <- readRevs gitDir (Ref sha)
Ref sha' <- lookupPath (BC.pack commit) cs
Commit (Just tree) _ _ <- readCommit gitDir $ Ref sha'
enter gitDir f "/" path tree
ls :: FilePath -> String -> Object -> IO ()
ls _ p o = case o of
Commit _ _ _ -> error "resolve to a commit"
Tree es -> mapM_ (BC.putStrLn . fst) $ treeToRefs es
Blob _ _ -> putStrLn p
cat :: FilePath -> String -> Object -> IO ()
cat _ _ o = case o of
Commit _ _ _ -> error "resolve to a commit"
Tree _ -> error "is a directory"
Blob _ bs -> L.putStr bs
export :: FilePath -> String -> Object -> IO ()
export gitDir p o = export' gitDir p ps o
where ps = if p == "/" then [] else [p]
export' :: FilePath -> String -> [String] -> Object -> IO ()
export' gitDir p ps o = case o of
Commit _ _ _ -> error "resolve to a commit"
Tree es -> do
let refs = treeToRefs es
f (p', ref) = do
o <- readObject gitDir ref
export' gitDir (BC.unpack p') (BC.unpack p':ps) o
mapM_ f refs
-- Blob _ bs -> putStrLn . ("Exporting " ++) . joinPath $ reverse ps
Blob _ _ -> L.putStr . runPut $ fileModify (BC.pack . joinPath $ reverse ps, o)
enter :: FilePath -> (FilePath -> String -> Object -> IO a) -> String -> [String] -> Ref -> IO a
enter gitDir f p ps ref = do
o <- readObject gitDir ref
case ps of
p':ps' -> case o of
Blob _ _ -> error $ "no file '" ++ p' ++ "'"
Tree es -> do
ref' <- lookupPath (BC.pack p') $ treeToRefs es
enter gitDir f p' ps' ref'
Commit _ _ _ -> error "Deref the tree ?"
[] -> f gitDir p o
lookupPath :: ByteString -> [(ByteString, a)] -> IO a
lookupPath k es = do
case lookup k es of
Just v -> return v
_ -> error $ "no file '" ++ BC.unpack k ++ "'" -- TODO exitFailure
-- TODO Also make it possible to specify the mode of each entry.
rewrite gitDir packer xs mref = do
o <- maybe (return $ Tree []) (readObject gitDir) mref
case o of
Blob _ _ -> error $ "file exists"
Tree es -> do
es' <- forM xs $ \x -> do
case x of
L name blob -> do
sha <- pack packer blob
return (normalFile, name, sha)
T name ys -> do
sha <- rewrite gitDir packer ys $ lookup name $ treeToRefs es
return (subdirectory, name, sha)
pack packer $ Tree . nubBy (\(_, a, _) (_, b, _) -> a == b) $ es' ++ es
Commit _ _ _ -> error "expected tree is a commit"
data T p a = T p [T p a] | L p a
deriving Show
input1 =
[ ([], "a", 1)
]
input2 =
[ ([], "a", 1)
, ([], "b", 2)
]
input3 =
[ ([], "a", 1)
, ([], "b", 2)
, (["e"], "c", 3)
]
input4 =
[ ([], "a", 1)
, ([], "b", 2)
, (["e"], "c", 3)
, (["f"], "d", 4)
]
input5 =
[ ([], "a", 1)
, ([], "b", 2)
, (["e"], "c", 3)
, (["f", "g"], "d", 4)
, (["f"], "i", 6)
, ([], "h", 5)
]
listBlobs = map listBlob . reverse . nubBy f . reverse
where f (a, _) (b, _) = a == b
listBlob (path, blob) = case splitDirectories $ BC.unpack path of
[] -> error "at least a filename must be given"
xs -> (map BC.pack $ init xs, BC.pack $ last xs, blob)
-- | Group blobs into a tree.
groupBlobs = groupBlobs' . listBlobs
groupBlobs' blobs = map (\(_, a, b) -> L a b) direct ++ rest
where
rest = map (\(d, bs) -> T d $ groupBlobs' bs) $ map pops $ groupBy f indirect
pops es@((x:_, _, _):_) = (x, map pop es)
pops [] = error "can't happen" -- groupBy returns non-empty lists
pop (_:xs, b, c) = (xs, b, c)
pop _ = error "can't happen" -- pop is called only on the indirect elements
(direct, indirect) = span isDirect $ sortBlobs blobs
isDirect ([], _, _) = True
isDirect _ = False
f ([], _, _) ([], _, _) = True -- unused, f is called on the indirect elements
f (x:_, _, _) (y:_, _, _) = x == y
-- | This is used to group blobs by path, only to arrange them in trees within
-- `groupBlobs`. The order is not the same as Git's Tree object. I.e. objects
-- will be written in the packfile in a slightly different order than they are
-- referenced in the Tree object.
sortBlobs :: Ord p => [([p], p, a)] -> [([p], p, a)]
sortBlobs = sortBy f
where
f (ps1, n1, _) (ps2, n2, _) = case compare ps1 ps2 of
EQ -> compare n1 n2
x -> x
-- | `repack _ fn` creates a new packfile stored at `fn` containg a tree of blobs
-- shadowing an optional tree. The resulting is referenced by commit, which
-- can receive an optional parent.
repack gitDir fn blobs msha mtree msg branch = do
packer <- newPack fn
tree' <- rewrite gitDir packer blobs mtree
Ref sha' <- pack packer $ Commit (Just tree') msha msg
completePack packer
B.writeFile branch $ sha' `BC.append` "\n" -- TODO Use git-update-ref.
return (Ref sha', tree')
----------------------------------------------------------------------
-- Read git objects and refs
----------------------------------------------------------------------
-- | Convert a Tree to the type returned by readHeads.
treeToRefs :: [(ByteString, ByteString, Ref)] -> [(ByteString, Ref)]
treeToRefs = map (\(_, b, c) -> (b, c))
newtype Sha = Sha ByteString
deriving (Eq, Show)
newtype Ref = Ref { unRef ::ByteString } -- TODO Can it actually be non-ascii ?
deriving Show
data Object =
Blob Int64 L.ByteString
| Tree [(ByteString, ByteString, Ref)] -- ^ Mode, name, sha
| Commit (Maybe Ref) (Maybe Ref) ByteString -- ^ Tree ref, parent ref, message.
deriving Show
-- | `git show-ref`
readRefs :: FilePath -> Maybe Ref -> IO [(Ref, Sha)]
readRefs gitDir mref = do
(code, out, _) <- readProcessWithExitCode' "git"
([ "show-ref", "--" ] ++ maybe [] s mref)
[("GIT_DIR", gitDir)]
""
if code == ExitSuccess
then return . map (p . words) $ lines out
-- git show-ref returns a exit code 1 when there is no ref.
-- TODO Differentiate between no ref and other non-zero exit codes.
else return []
where s (Ref r) =[ BC.unpack r]
p [sha, r] = (Ref $ BC.pack r, Sha $ BC.pack sha)
p _ = error "unexpected git-show-ref output"
-- | Like readRefs, but return only those matching `refs/heads`.
readHeads :: FilePath -> Maybe Ref -> IO [(ByteString, Sha)]
readHeads gitDir mref = do
refs <- readRefs gitDir mref
return $ map unref $ filter (prefix . fst) refs
where unref (Ref r, sha) = (BC.drop 11 r, sha)
prefix (Ref r) = BC.isPrefixOf "refs/heads/" r
-- | `git cat-file --batch`
-- TODO Keep the process `git cat-file` around and query it instead of
-- respawning it again and again.
readObjects :: FilePath -> [Ref] -> IO [Object]
readObjects gitDir refs = do
(Just pIn, Just pOut, _, p) <- createProcess (proc "git"
[ "cat-file", "--batch"
])
{ std_in = CreatePipe
, std_out = CreatePipe
, env = Just [("GIT_DIR", gitDir)]
}
let putRef (Ref r) = do
BC.hPutStrLn pIn r
hFlush pIn
readOne = do
ws <- BC.words <$> BC.hGetLine pOut
case ws of
[sha, typ, size_] | Just (size, _) <- BC.readInteger size_ -> do
-- TODO hGet takes an Int, maybe we should read again if the Int64
-- is really useful.
o <- L.hGet pOut (fromInteger size)
nl <- BC.hGetLine pOut
when (nl /= "") $ error "unexpected git-cat-file output (1)"
case typ of
"blob" -> return $ Blob (fromInteger size) o
"tree" -> do
let loop xs s = do
if L.null s
then return . Tree $ reverse xs
else do
-- Maybe rewrite this with attoparsec.
let (a, b) = LC.span (/= ' ') s
c = LC.drop 1 b
(d, d') = LC.span (/= '\0') c
e = LC.drop 1 d'
(f, g) = LC.splitAt 20 e
loop ((toStrict a,toStrict d, Ref . B16.encode $ toStrict f): xs) g
loop [] o
"commit" -> return $ parseCommit o
_ -> error "unexpected git-cat-file output (2)"
x -> error $ "unexpected git-cat-file output (3)" ++ show x
os <- mapM (\r -> putRef r >> readOne) refs
hClose pIn
_ <- waitForProcess p
return os
toStrict :: L.ByteString -> ByteString
toStrict = B.concat . L.toChunks
-- | Similar to `readObjects` with a single ref.
readObject :: FilePath -> Ref -> IO Object
readObject gitDir ref = do
os <- readObjects gitDir [ref]
case os of
[o] -> return o
_ -> error $ "can't happen"
-- | Similar to `readObjects` (with a single ref), and error out if the result
-- is not a blob.
readBlob :: FilePath -> Ref -> IO Object
readBlob gitDir ref = do
o <- readObject gitDir ref
case o of
Blob _ _ -> return o
_ -> error $ "expected blob object"
-- | Similar to `readObjects` (with a single ref), and error out if the result
-- is not a commit.
readCommit :: FilePath -> Ref -> IO Object
readCommit gitDir ref = do
o <- readObject gitDir ref
case o of
Commit _ _ _ -> return o
_ -> error $ "expected commit object"
-- | Similar to `readObjects` (with a single ref), and error out if the result
-- is not a tree.
readTree :: FilePath -> Ref -> IO Object
readTree gitDir ref = do
o <- readObject gitDir ref
case o of
Tree _ -> return o
_ -> error $ "expected tree object"
parseCommit :: L.ByteString -> Object
parseCommit bs = case parse p bs of
Fail _ _ err -> error err
Done _ r -> r
where
p = do
_ <- A.string "tree " -- TODO Optional.
treeSha <- A.takeWhile isSha
when (B.length treeSha /= 40) $ error "unexpected tree ref length"
_ <- A.char '\n'
-- TODO
_ <- A.takeByteString
A.endOfInput
return $ Commit (Just $ Ref treeSha) Nothing ""
isSha c = (c >= '0' && c <= '9') ||
(c >= 'a' && c <= 'f') ||
(c >= 'A' && c <= 'F')
-- | `git rev-list --pretty=format:%at`
revList :: FilePath -> Ref -> IO [(Ref, UTCTime)]
revList gitDir ref = do
(code, out, _) <- readProcessWithExitCode' "git"
[ "rev-list", "--pretty=format:%at", BC.unpack $ unRef ref ]
[("GIT_DIR", gitDir)]
""
if code == ExitSuccess
then return . p [] $ lines out
else error "git failed"
where
-- TODO read
p xs (l1:l2:rest) | "commit":_ <- words l2 =
p ((Ref . BC.pack $ drop 7 l1, posixSecondsToUTCTime . fromInteger $ 0) : xs) (l2:rest)
p xs (l1:l2:rest) =
p ((Ref . BC.pack $ drop 7 l1, posixSecondsToUTCTime . fromInteger $ read l2) : xs) rest
p xs [l1] | "commit":_ <- words l1 =
p ((Ref . BC.pack $ drop 7 l1, posixSecondsToUTCTime . fromInteger $ 0) : xs) []
p xs [] = reverse xs
p _ _ = error "unexpected line from git-rev-list"
-- | Similar to `revList` but the result type matches `readHeads`.
readRevs :: FilePath -> Ref -> IO [(ByteString, Ref)]
readRevs gitDir ref = do
refs <- revList gitDir ref
return . latest $ map f refs
where
f (r, t) = (BC.pack $ formatTime locale format t, r)
locale = defaultTimeLocale
format = "%Y-%m-%d-%H%M%S"
latest (x@(_, r) : xs) = ("latest", r) : x : xs
latest _ = []
----------------------------------------------------------------------
-- Write packfile
--
-- A packfile can be verified with `git verify-pack`. It needs a corresponding
-- `.idx` file which be can generated with `git index-pack`. E.g.:
--
-- > git index-pack t.pack
-- > git verify-pack -v t.pack
-- 32ab47487f560b11fdc758eedd9b43ee7aeb8684 blob 749 349 12
-- non delta: 1 object
-- t.pack: ok
--
-- Those two commands don't need a a Git repository to work. On the other
-- hand, to use a command such as `git cat-file`, a real Git repository must
-- be provided:
--
-- > git init --bare repo
-- > cp t.{idx,pack} repo/objects/pack/
-- > cd repo
-- > echo 32ab47487f560b11fdc758eedd9b43ee7aeb8684 | git cat-file --batch
-- 32ab47487f560b11fdc758eedd9b43ee7aeb8684 blob 749
-- ... blob content ...
--
-- If the packfile contain a commit, we can pretend HEAD points to it, inspect
-- it, or even do a checkout:
--
-- > echo 709149cd69d4e13c8740e5bb3d832f97fcb08878 > refs/heads/master
-- > git log
-- commit 709149cd69d4e13c8740e5bb3d832f97fcb08878
--
-- > mkdir ../work
-- > GIT_WORK_TREE=../work git checkout master
-- Already on 'master'
--
----------------------------------------------------------------------
-- | Incrementally build a pack. It also builds the index. TODO Build the index
-- as the packfile is built, not afterwards.
newPack :: FilePath -> IO Packer
newPack fn = do
h <- openFile fn ReadWriteMode
hSetFileSize h 0 -- TODO Instead use a temporary (and thus new) file,
-- moving it to the correct path when it is complete.
-- The number of objects will be set in `completePack`.
BC.hPut h "PACK\0\0\0\2\0\0\0\0"
counter <- newIORef 0
return Packer
{ packerPack = \o -> do
modifyIORef counter succ
let (sha, bs) = runPutM $ putObject o
L.hPut h bs
return sha
, packerComplete = do
hSeek h AbsoluteSeek 8
n <- readIORef counter
L.hPut h . runPut . putWord32be $ n
hSeek h AbsoluteSeek 0
content <- hReadAll h
let sha = bytestringDigest $ sha1 content
L.hPut h sha
hClose h
indexPack fn
}
indexPack :: String -> IO ()
indexPack path = do
(_, _, _, p) <- createProcess (proc "git"
["index-pack", "-v", path])
_ <- waitForProcess p
return ()
-- | This is the function hGetContentsN from the bytestring package, minus the
-- handle closing bit of code.
hReadAll :: Handle -> IO L.ByteString
hReadAll h = lazyRead -- TODO close on exceptions
where
lazyRead = unsafeInterleaveIO loop
loop = do
c <- B.hGetSome h I.defaultChunkSize -- only blocks if there is no data available
if B.null c
then return I.Empty
else do
cs <- lazyRead
return $ I.Chunk c cs
pack :: Packer -> Object -> IO Ref
pack packer = packerPack packer
pack_ :: Packer -> Object -> IO ()
pack_ packer o = packerPack packer o >> return ()
completePack :: Packer -> IO ()
completePack = packerComplete
data Packer = Packer
{ packerPack :: Object -> IO Ref
, packerComplete :: IO ()
}
-- | Write a packfile. The content of the packfile is provided as a
-- `Data.Binary.Put` serializer. The number of objects must be provided
-- explicitely.
writePack :: FilePath -> Int -> Put -> IO ()
writePack fn n os = L.writeFile fn p
-- TODO Compute the SHA1 sum on-the-fly.
where p_ = runPut $ buildPack n os
sha = bytestringDigest $ sha1 p_
p = p_ `L.append` sha
-- | Build a packfile, minus its SHA1 sum.
buildPack :: Int -> Put -> Put
buildPack n os = do
putByteString "PACK\0\0\0\2"
putWord32be . fromIntegral $ n
os
-- | Serialize an object, using the packfile format.
putObject :: Object -> PutM Ref
putObject o = case o of
Blob size bs -> do
putLength 3 size -- Assume that L.length bs == size.
putLazyByteString $ compress bs
return $ computeSha o
Tree es -> do
let bs = runPut $ putTree es
putLength 2 $ L.length bs
putLazyByteString $ compress bs
return $ computeSha o
Commit mtree mparent msg -> do
let bs = runPut $ putCommit mtree mparent Nothing Nothing msg
putLength 1 $ L.length bs
putLazyByteString $ compress bs
return $ computeSha o
-- | Each object stored in a packfile still retain its loose object SHA1 sum.
computeSha :: Object -> Ref
computeSha o =
Ref . B16.encode . toStrict . bytestringDigest . sha1 . runPut $ putLoose o
-- | Serialize an object using the loose format (but not yet zlib compressed).
putLoose :: Object -> Put
putLoose o = case o of
Blob size bs -> do
putByteString "blob "
putByteString (BC.pack $ show size)
putWord8 0
putLazyByteString bs
Tree es -> do
let bs = runPut $ putTree es
putByteString "tree "
putByteString (BC.pack $ show $ L.length bs)
putWord8 0
putLazyByteString bs
Commit mtree mparent msg -> do
let bs = runPut $ putCommit mtree mparent Nothing Nothing msg
putByteString "commit "
putByteString (BC.pack $ show $ L.length bs)
putWord8 0
putLazyByteString bs
-- | Variable length unsigned integer encoding, used in the packfile format.
-- The type of the object is included.
putLength :: Word8 -> Int64 -> Put
putLength t n = loop size b
where
-- Object type is in the three last bits of the first nibble
-- The first bit (not yet set) is the "continue" bit.
-- / / The second nibble contains the size.
b = (shiftL t 4) .|. (fromIntegral n .&. 0x0f)
size = shiftR n 4
loop sz c =
if sz /= 0
then do
putWord8 $ c .|. 0x80 -- set the "continue"
loop (shiftR sz 7) (fromIntegral sz .&. 0x7f) -- and continue with the next 7 bits
else putWord8 c
-- | Write triple (Mode, name, sha) as a `tree` object in the packfile format.
putTree :: [(ByteString, ByteString, Ref)] -> Put
putTree es = mapM_ putEntry es'
where
es' = sortBy filenames es
filenames (mode1, n1, _) (mode2, n2, _) = compare (f mode1 n1) (f mode2 n2)
where f mode n = if mode == subdirectory then n `B.append` "/" else n
putEntry (mode, name, Ref sha) = do
putByteString mode
putWord8 32 -- that is ' '
putByteString name
putWord8 0
case B16.decode sha of
(sha', rest) | B.null rest -> putByteString sha'
_ -> error "putEntry: invalid sha"
putCommit :: Maybe Ref -> Maybe Ref -> Maybe (ByteString, UTCTime)
-> Maybe (ByteString, UTCTime) -> ByteString -> Put
putCommit mtree mparent mauthor mcommitter msg = do
let opt s f m = do
maybe (return ()) (\v -> do
putByteString s
putByteString $ f v
putWord8 10 -- that is '\n'
) m
opt "tree " unRef mtree
opt "parent " unRef mparent
-- TODO
putWord8 10
putByteString msg
normalFile :: ByteString
normalFile = "100644"
subdirectory :: ByteString
subdirectory = "040000"
----------------------------------------------------------------------
-- serialization to `git fast-import` format
--
-- Example usage:
-- buh fast-export | git fast-import --date-format=now
----------------------------------------------------------------------
fastExport gitDir mfrom files = L.putStr $ toFastExport mfrom files
toFastExport mfrom files = runPut (feCommit mfrom >> feFiles files)
feCommit (mfrom) = do
putByteString "commit refs/heads/fast-import\n"
-- mark?
-- author?
putByteString "committer Vo Minh Thu <[email protected]> now\n" -- TODO git date format
putByteString "data 0\n"
maybe (return ()) (\ref -> putByteString $ B.concat ["from ", ref, "^0\n"]) mfrom
feFiles = mapM_ fileModify
fileModify (path, Blob n bs) = do
putByteString "M "
putByteString normalFile
putByteString " inline "
putByteString path
putByteString "\ndata "
putByteString . BC.pack $ show n
putByteString "\n"
putLazyByteString bs
putByteString "\n"
-- | Same as System.Process.readProcessWithExitCode but allow to pass an
-- environment.
readProcessWithExitCode'
:: FilePath -- ^ command to run
-> [String] -- ^ any arguments
-> [(String, String)] -- ^ environment
-> String -- ^ standard input
-> IO (ExitCode,String,String) -- ^ exitcode, stdout, stderr
readProcessWithExitCode' cmd args env input = do
(Just inh, Just outh, Just errh, pid) <-
createProcess (proc cmd args)
{ std_in = CreatePipe
, std_out = CreatePipe
, std_err = CreatePipe
, env = Just env
}
outMVar <- newEmptyMVar
-- fork off a thread to start consuming stdout
out <- hGetContents outh
forkIO $ C.evaluate (length out) >> putMVar outMVar ()
-- fork off a thread to start consuming stderr
err <- hGetContents errh
forkIO $ C.evaluate (length err) >> putMVar outMVar ()
-- now write and flush any input
when (not (null input)) $ do hPutStr inh input; hFlush inh
hClose inh -- done with stdin
-- wait on the output
takeMVar outMVar
takeMVar outMVar
hClose outh
-- wait on the process
ex <- waitForProcess pid
return (ex, out, err)
| noteed/buh | bin/buh.hs | bsd-3-clause | 28,165 | 0 | 37 | 7,141 | 8,496 | 4,331 | 4,165 | 605 | 14 |
-----------------------------------------------------------------------------
-- |
-- Copyright : (C) 2015 Dimitri Sabadie
-- License : BSD3
--
-- Maintainer : Dimitri Sabadie <[email protected]>
-- Stability : experimental
-- Portability : portable
--
-- Position in space is a 3-float vector.
----------------------------------------------------------------------------
module Quaazar.Geometry.Position (
-- * Position
Position(..)
, pos
) where
import Data.Aeson
import Data.Aeson.Types ( typeMismatch )
import Linear ( V3(..) )
import Quaazar.Render.GL.Shader ( Uniformable(..) )
-- |Position in space a.k.a. space coordinates.
newtype Position = Position { unPosition :: V3 Float } deriving (Eq,Ord,Show)
instance FromJSON Position where
parseJSON v = do
a <- parseJSON v
case a of
[x,y,z] -> return (pos x y z)
_ -> typeMismatch "position" v
instance Uniformable Position where
sendUniform l = sendUniform l . unPosition
-- |Build a 'Position' from /x/, /y/ and /z/ components.
pos :: Float -> Float -> Float -> Position
pos x y z = Position (V3 x y z)
| phaazon/quaazar | src/Quaazar/Geometry/Position.hs | bsd-3-clause | 1,124 | 0 | 13 | 210 | 245 | 141 | 104 | 18 | 1 |
module Tutorial where
import MFlow.Wai.Blaze.Html.All
import Data.Monoid
-- widget signature : View rendering monad returnValue
-- flow aignature: FlowM rendering monad returnValue
-- page : View v m a -> FlowM v m a
main = runNavigation "" . step $ do
r <- page $ h3 << "Basic example with inputs and links"
++> getString (Just "text input")
<* submitButton "OK"
page $ b << ("you typed: "++ r)
++> p << "wlink's return typed values to the flow"
++> p << "in MFlow a page is a widget made of widgets"
++> wlink () << p << "next"
r <- page $ h3 << "Operators"
++> p << " The operator (++>) prepend HTML to a widget"
++> p << "in this case, a field that expect an Int"
++> getInt Nothing
<! [("placeholder","This is an attribute for the getInt form")]
<* submitButton "OK"
<++ p << "the operator (<<) add text to a blaze-html tag in this case"
<> p << "If the OverloadedStrings extension is used, this is not necessary"
<> p << " Note that form tags are added automatically"
page $ b << ("you entered: "++ show r) ++> wlink () << p << "next"
r <- page $ h3 << "Here the alternative operator is used to choose between two options"
++> wlink True << b << "True" <++ br
<|> wlink False << b << "False"
page $ b << ("you entered: "++ show r) ++> wlink () << p << "next"
-- |+| <+>
r <- page $ ul << h3 << "More operators"
++> li << " (<<<) embed a widget in a tag"
++> li << " (|*>) intersperse a widget within a list of others"
++> ul << li << "in this case, a link is interspersed within a list of input fields"
++> ((h3 <<< wlink "next" << b << "next")
|*> [getString Nothing <![("placeholder","enter field "++ show i)] | i <- [1..3]])
<** submitButton "OK"
case r of
(Just _, Nothing) -> return ()
(Nothing, Just s) -> do
page $ p << ("you entered " ++ s ++ " in some box")
++> wlink " next" << b << "next"
return()
page $ pageFlow "first" $ do
r <- h3 << "Page flows: run within the View monad, within a page"
++> p << "The call \"pageFlow\" creates a set of unique identifiers and stores\
\the data entered during the steps of the pageflow to replay the computation\
\each time the page is refreshed."
++> p << "until the last monadic statement does not validate, the page Flow will execute again and again"
++> wlink True << b << "True" <++ br
<|> wlink False << b << "False"
p << "Until the first line is not validated, the second does not execute"
++> p << ("you entered: "++ show r) ++> wlink () << p << "click"
page $ pageFlow "second" $ do
h2 << "Field validators" ++> wlink () << p << "next"
r <- p << "this field expect a string less than 5 characters. Otherwise it present a alert message"
++> getString Nothing
`validate`
(\r -> if length r > 5 then return . Just $ script << "alert ('length must be lsst than five chars ')"
else return Nothing)
<* submitButton "OK"
p << ("you entered: "++ show r) ++> wlink () << p << "click"
page $ pageFlow "third" $ do
h2 << "A select/option box" ++> wlink () << b << "click" <++ br
r <- getSelect
(setSelectedOption "" << p << "select a option" <|>
setOption "red" << b << "red" <|>
setOption "blue" << b << "blue" <|>
setOption "Green" << b << "Green")
<! [("onchange","this.form.submit()")]
p << (r ++ " selected") ++> wlink () << p << "next"
let colors= getCheckBoxes( (setCheckBox False "Red" <++ b << "red")
<> (setCheckBox False "Green" <++ b << "green")
<> (setCheckBox False "blue" <++ b << "blue"))
page $ pageFlow "four" $ do
h2 << "checkboxes" ++> wlink () << b << "click" <++ br
r <- colors <** submitButton "submit"
p << (show r ++ " selected") ++> wlink () << p << " menu"
page $ (do
h3 << "A mix of applicative and monadic operators can be used to create multiple pageflows within a page"
++> p << "here two pageflows are composed with an alternative operator"
++> p << "the button triggers the presentation of the changes in both elements."
++> p << "they are part of the same HTML form, but they change their display depending on the input"
++> wlink "" << p << "click here"
(pageFlow "colors" (do
r <- colors <++ br
p << (show r ++ "selected") ++> noWidget <++ br)
<|>
pageFlow "input" (do
r <- getString Nothing <++ br
p << (show r ++"entered") ++> noWidget <++ br)
<* submitButton "OK" ))
<|> br ++> wlink "next" << b << "skip to the next"
<++ p << "this link is alternative (<|>) to the rest of the page, so it can be pressed\
\to skip it at any moment"
return ()
main= runNavigation "" .step . page (pageFlow "s" ( do
n <- getInt Nothing <++ "first"
n' <- getInt (Just n) <++ "second"
p << (n+n') ++> wlink () "click to repeat")
<** submitButton "Send")
| agocorona/MFlow | Demos/Tutorial 1.hs | bsd-3-clause | 5,648 | 1 | 30 | 2,011 | 1,408 | 681 | 727 | 93 | 3 |
module Blog.BackEnd.RefererStream where
import qualified System.Log.Logger as L
import qualified Blog.FrontEnd.Views as V
import qualified Blog.Constants as C
import qualified Control.Monad as CM
import qualified Data.Map as DM
import Data.List ( isPrefixOf )
import Control.Concurrent
import Control.Concurrent.Chan
import Control.Concurrent.MVar
data RefererStream = RefererStream { control :: Chan Request}
data Request = AddReferer { view_url :: String, referring_url :: String }
| GetReferers { handback :: MVar Referers }
boot :: IO RefererStream
boot = do { c <- newChan
; let rs = RefererStream c
; forkIO $ referer_loop rs empty_referers
; return rs }
type Referers = DM.Map String (DM.Map String Int)
log_handle :: String
log_handle = "RefererStream"
empty_referers :: Referers
empty_referers = DM.empty
send_referer :: (V.Viewable v) => RefererStream -> v -> String -> IO ()
send_referer rs view e = writeChan ( control rs ) $ AddReferer (V.url view) e
get_referers :: (V.Viewable v) => RefererStream -> v -> IO (DM.Map String Int)
get_referers rs v = do { h <- newEmptyMVar
; writeChan ( control rs) $ GetReferers h
; r <- (takeMVar h)
; return $ DM.findWithDefault DM.empty (V.url v) r }
add_referer :: Request -> Referers -> Referers
add_referer (AddReferer v r) m | C.blog_root `isPrefixOf` r = m
| v `DM.member` m = DM.adjust (DM.insertWith' (+) r 1) v m
| otherwise = DM.insert v (DM.insert r 1 DM.empty) m
referer_loop :: RefererStream -> Referers -> IO ()
referer_loop rs r = do { req <- readChan . control $ rs
; L.infoM log_handle $ show r
; case req of
AddReferer _ _ ->
referer_loop rs $ add_referer req r
GetReferers h ->
putMVar h r >> referer_loop rs r }
create_referrers_tables_sql :: String
create_referrers_tables_sql =
"CREATE TABLE referers ( permatitle TEXT PRIMARY KEY NOT NULL, "
++ "referring_uri TEXT NOT NULL, "
++ "first_hit INTEGER NOT NULL, "
++ "most_recent_hit INTEGER NOT NULL, "
++ "count INTEGER NOT NULL DEFAULT 0 )"
| prb/perpubplat | src/Blog/BackEnd/RefererStream.hs | bsd-3-clause | 2,355 | 3 | 12 | 713 | 672 | 355 | 317 | 49 | 2 |
{-# LANGUAGE BangPatterns #-}
{-# LANGUAGE CPP #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE GADTs #-}
{-# LANGUAGE IncoherentInstances #-}
{-# LANGUAGE NoForeignFunctionInterface #-}
{-# LANGUAGE PatternGuards #-}
{-# LANGUAGE ScopedTypeVariables #-}
{-# LANGUAGE TemplateHaskell #-}
{-# LANGUAGE TypeOperators #-}
{-# LANGUAGE TypeSynonymInstances #-}
{-# LANGUAGE UndecidableInstances #-}
-- |
-- Module : Data.Array.Accelerate.CUDA.Execute
-- Copyright : [2008..2014] Manuel M T Chakravarty, Gabriele Keller
-- [2009..2014] Trevor L. McDonell
-- License : BSD3
--
-- Maintainer : Trevor L. McDonell <[email protected]>
-- Stability : experimental
-- Portability : non-portable (GHC extensions)
--
module Data.Array.Accelerate.CUDA.Execute (
-- * Execute a computation under a CUDA environment
executeAcc, executeAfun1
) where
-- friends
import Data.Array.Accelerate.CUDA.AST
import Data.Array.Accelerate.CUDA.State
import Data.Array.Accelerate.CUDA.FullList ( FullList(..), List(..) )
import Data.Array.Accelerate.CUDA.Array.Data
import Data.Array.Accelerate.CUDA.Array.Sugar
import Data.Array.Accelerate.CUDA.Foreign.Import ( canExecuteAcc )
import Data.Array.Accelerate.CUDA.CodeGen.Base ( Name, namesOfArray, groupOfInt )
import Data.Array.Accelerate.CUDA.Execute.Event ( Event )
import Data.Array.Accelerate.CUDA.Execute.Stream ( Stream )
import qualified Data.Array.Accelerate.CUDA.Array.Prim as Prim
import qualified Data.Array.Accelerate.CUDA.Debug as D
import qualified Data.Array.Accelerate.CUDA.Execute.Event as Event
import qualified Data.Array.Accelerate.CUDA.Execute.Stream as Stream
import Data.Array.Accelerate.Error
import Data.Array.Accelerate.Tuple
import Data.Array.Accelerate.Interpreter ( evalPrim, evalPrimConst, evalPrj )
import Data.Array.Accelerate.Array.Data ( ArrayElt, ArrayData )
import Data.Array.Accelerate.Array.Representation ( SliceIndex(..) )
import qualified Data.Array.Accelerate.Array.Representation as R
-- standard library
import Prelude hiding ( exp, sum, iterate )
import Control.Applicative hiding ( Const )
import Control.Monad ( join, when, liftM )
import Control.Monad.Reader ( asks )
import Control.Monad.State ( gets )
import Control.Monad.Trans ( MonadIO, liftIO )
import System.IO.Unsafe ( unsafeInterleaveIO )
import Data.Int
import Data.Word
import Data.Maybe
import Foreign.CUDA.Analysis.Device ( computeCapability, Compute(..) )
import qualified Foreign.CUDA.Driver as CUDA
import qualified Data.HashMap.Strict as Map
-- Asynchronous kernel execution
-- -----------------------------
-- Arrays with an associated CUDA Event that will be signalled once the
-- computation has completed.
--
data Async a = Async {-# UNPACK #-} !Event !a
-- Valuation for an environment of asynchronous array computations
--
data Aval env where
Aempty :: Aval ()
Apush :: Aval env -> Async t -> Aval (env, t)
-- Projection of a value from a valuation using a de Bruijn index.
--
aprj :: Idx env t -> Aval env -> Async t
aprj ZeroIdx (Apush _ x) = x
aprj (SuccIdx idx) (Apush val _) = aprj idx val
aprj _ _ = $internalError "aprj" "inconsistent valuation"
-- All work submitted to the given stream will occur after the asynchronous
-- event for the given array has been fulfilled. Synchronisation is performed
-- efficiently on the device. This function returns immediately.
--
after :: MonadIO m => Stream -> Async a -> m a
after stream (Async event arr) = liftIO $ Event.after event stream >> return arr
-- Block the calling thread until the event for the given array computation
-- is recorded.
--
wait :: MonadIO m => Async a -> m a
wait (Async e x) = liftIO $ Event.block e >> return x
-- Execute the given computation in a unique execution stream.
--
streaming :: (Stream -> CIO a) -> (Async a -> CIO b) -> CIO b
streaming first second = do
context <- asks activeContext
reservoir <- gets streamReservoir
Stream.streaming context reservoir first (\e a -> second (Async e a))
-- Array expression evaluation
-- ---------------------------
-- Computations are evaluated by traversing the AST bottom-up, and for each node
-- distinguishing between three cases:
--
-- 1. If it is a Use node, return a reference to the device memory holding the
-- array data
--
-- 2. If it is a non-skeleton node, such as a let-binding or shape conversion,
-- this is executed directly by updating the environment or similar
--
-- 3. If it is a skeleton node, the associated binary object is retrieved,
-- memory allocated for the result, and the kernel(s) that implement the
-- skeleton are invoked
--
executeAcc :: Arrays a => ExecAcc a -> CIO a
executeAcc !acc = streaming (executeOpenAcc acc Aempty) wait
executeAfun1 :: (Arrays a, Arrays b) => ExecAfun (a -> b) -> a -> CIO b
executeAfun1 !afun !arrs = do
streaming (useArrays (arrays arrs) (fromArr arrs))
(\(Async event ()) -> executeOpenAfun1 afun Aempty (Async event arrs))
where
useArrays :: ArraysR arrs -> arrs -> Stream -> CIO ()
useArrays ArraysRunit () _ = return ()
useArrays (ArraysRpair r1 r0) (a1, a0) st = useArrays r1 a1 st >> useArrays r0 a0 st
useArrays ArraysRarray arr st = useArrayAsync arr (Just st)
executeOpenAfun1 :: PreOpenAfun ExecOpenAcc aenv (a -> b) -> Aval aenv -> Async a -> CIO b
executeOpenAfun1 (Alam (Abody f)) aenv x = streaming (executeOpenAcc f (aenv `Apush` x)) wait
executeOpenAfun1 _ _ _ = error "the sword comes out after you swallow it, right?"
-- Evaluate an open array computation
--
executeOpenAcc
:: forall aenv arrs.
ExecOpenAcc aenv arrs
-> Aval aenv
-> Stream
-> CIO arrs
executeOpenAcc EmbedAcc{} _ _
= $internalError "execute" "unexpected delayed array"
executeOpenAcc (ExecAcc (FL () kernel more) !gamma !pacc) !aenv !stream
= case pacc of
-- Array introduction
Use arr -> return (toArr arr)
Unit x -> newArray Z . const =<< travE x
-- Environment manipulation
Avar ix -> after stream (aprj ix aenv)
Alet bnd body -> streaming (executeOpenAcc bnd aenv) (\x -> executeOpenAcc body (aenv `Apush` x) stream)
Apply f a -> streaming (executeOpenAcc a aenv) (executeOpenAfun1 f aenv)
Atuple tup -> toTuple <$> travT tup
Aprj ix tup -> evalPrj ix . fromTuple <$> travA tup
Acond p t e -> travE p >>= \x -> if x then travA t else travA e
Awhile p f a -> awhile p f =<< travA a
-- Foreign
Aforeign ff afun a -> fromMaybe (executeAfun1 afun) (canExecuteAcc ff) =<< travA a
-- Producers
Map _ a -> executeOp =<< extent a
Generate sh _ -> executeOp =<< travE sh
Transform sh _ _ _ -> executeOp =<< travE sh
Backpermute sh _ _ -> executeOp =<< travE sh
Reshape sh a -> reshapeOp <$> travE sh <*> travA a
-- Consumers
Fold _ _ a -> foldOp =<< extent a
Fold1 _ a -> fold1Op =<< extent a
FoldSeg _ _ a s -> join $ foldSegOp <$> extent a <*> extent s
Fold1Seg _ a s -> join $ foldSegOp <$> extent a <*> extent s
Scanl1 _ a -> scan1Op =<< extent a
Scanr1 _ a -> scan1Op =<< extent a
Scanl' _ _ a -> scan'Op =<< extent a
Scanr' _ _ a -> scan'Op =<< extent a
Scanl _ _ a -> scanOp True =<< extent a
Scanr _ _ a -> scanOp False =<< extent a
Permute _ d _ a -> join $ permuteOp <$> extent a <*> travA d
Stencil _ _ a -> stencilOp =<< travA a
Stencil2 _ _ a1 _ a2 -> join $ stencil2Op <$> travA a1 <*> travA a2
-- Removed by fusion
Replicate _ _ _ -> fusionError
Slice _ _ _ -> fusionError
ZipWith _ _ _ -> fusionError
where
fusionError = $internalError "executeOpenAcc" "unexpected fusible matter"
-- term traversals
travA :: ExecOpenAcc aenv a -> CIO a
travA !acc = executeOpenAcc acc aenv stream
travE :: ExecExp aenv t -> CIO t
travE !exp = executeExp exp aenv stream
travT :: Atuple (ExecOpenAcc aenv) t -> CIO t
travT NilAtup = return ()
travT (SnocAtup !t !a) = (,) <$> travT t <*> travA a
awhile :: PreOpenAfun ExecOpenAcc aenv (a -> Scalar Bool) -> PreOpenAfun ExecOpenAcc aenv (a -> a) -> a -> CIO a
awhile p f a = do
nop <- liftIO Event.create -- record event never call, so this is a functional no-op
r <- executeOpenAfun1 p aenv (Async nop a)
ok <- indexArray r 0 -- TLM TODO: memory manager should remember what is already on the host
if ok then awhile p f =<< executeOpenAfun1 f aenv (Async nop a)
else return a
-- get the extent of an embedded array
extent :: Shape sh => ExecOpenAcc aenv (Array sh e) -> CIO sh
extent ExecAcc{} = $internalError "executeOpenAcc" "expected delayed array"
extent (EmbedAcc sh) = travE sh
-- Skeleton implementation
-- -----------------------
-- Execute a skeleton that has no special requirements: thread decomposition
-- is based on the given shape.
--
executeOp :: (Shape sh, Elt e) => sh -> CIO (Array sh e)
executeOp !sh = do
out <- allocateArray sh
execute kernel gamma aenv (size sh) out stream
return out
-- Change the shape of an array without altering its contents. This does not
-- execute any kernel programs.
--
reshapeOp :: Shape sh => sh -> Array sh' e -> Array sh e
reshapeOp sh (Array sh' adata)
= $boundsCheck "reshape" "shape mismatch" (size sh == R.size sh')
$ Array (fromElt sh) adata
-- Executing fold operations depend on whether we are recursively collapsing
-- to a single value using multiple thread blocks, or a multidimensional
-- single-pass reduction where there is one block per inner dimension.
--
fold1Op :: (Shape sh, Elt e) => (sh :. Int) -> CIO (Array sh e)
fold1Op !sh@(_ :. sz)
= $boundsCheck "fold1" "empty array" (sz > 0)
$ foldCore sh
foldOp :: (Shape sh, Elt e) => (sh :. Int) -> CIO (Array sh e)
foldOp !(!sh :. sz)
= foldCore ((listToShape . map (max 1) . shapeToList $ sh) :. sz)
foldCore :: (Shape sh, Elt e) => (sh :. Int) -> CIO (Array sh e)
foldCore !(!sh :. sz)
| dim sh > 0 = executeOp sh
| otherwise
= let !numElements = size sh * sz
(_,!numBlocks,_) = configure kernel numElements
in do
out <- allocateArray (sh :. numBlocks)
execute kernel gamma aenv numElements out stream
foldRec out
-- Recursive step(s) of a multi-block reduction
--
foldRec :: (Shape sh, Elt e) => Array (sh:.Int) e -> CIO (Array sh e)
foldRec arr@(Array _ !adata)
| Cons _ rec _ <- more
= let sh :. sz = shape arr
!numElements = size sh * sz
(_,!numBlocks,_) = configure rec numElements
in if sz <= 1
then return $ Array (fromElt sh) adata
else do
out <- allocateArray (sh :. numBlocks)
execute rec gamma aenv numElements (out, arr) stream
foldRec out
| otherwise
= $internalError "foldRec" "missing phase-2 kernel module"
-- Segmented reduction. Subtract one from the size of the segments vector as
-- this is the result of an exclusive scan to calculate segment offsets.
--
foldSegOp :: (Shape sh, Elt e) => (sh :. Int) -> (Z :. Int) -> CIO (Array (sh :. Int) e)
foldSegOp (!sh :. _) !(Z :. sz) = executeOp (sh :. sz - 1)
-- Scans, all variations on a theme.
--
scanOp :: Elt e => Bool -> (Z :. Int) -> CIO (Vector e)
scanOp !left !(Z :. numElements) = do
arr@(Array _ adata) <- allocateArray (Z :. numElements + 1)
out <- devicePtrsOfArrayData adata
let (!body, !sum)
| left = (out, advancePtrsOfArrayData adata numElements out)
| otherwise = (advancePtrsOfArrayData adata 1 out, out)
--
scanCore numElements arr body sum
return arr
scan1Op :: forall e. Elt e => (Z :. Int) -> CIO (Vector e)
scan1Op !(Z :. numElements) = do
arr@(Array _ adata) <- allocateArray (Z :. numElements + 1) :: CIO (Vector e)
body <- devicePtrsOfArrayData adata
let sum {- to fix type -} = advancePtrsOfArrayData adata numElements body
--
scanCore numElements arr body sum
return (Array ((),numElements) adata)
scan'Op :: forall e. Elt e => (Z :. Int) -> CIO (Vector e, Scalar e)
scan'Op !(Z :. numElements) = do
vec@(Array _ ad_vec) <- allocateArray (Z :. numElements) :: CIO (Vector e)
sum@(Array _ ad_sum) <- allocateArray Z :: CIO (Scalar e)
d_vec <- devicePtrsOfArrayData ad_vec
d_sum <- devicePtrsOfArrayData ad_sum
--
scanCore numElements vec d_vec d_sum
return (vec, sum)
scanCore
:: forall e. Elt e
=> Int
-> Vector e -- to fix Elt vs. EltRepr
-> Prim.DevicePtrs (EltRepr e)
-> Prim.DevicePtrs (EltRepr e)
-> CIO ()
scanCore !numElements (Array _ !adata) !body !sum
| Cons _ !upsweep1 (Cons _ !upsweep2 _) <- more
= let (_,!numIntervals,_) = configure kernel numElements
!d_body = marshalDevicePtrs adata body
!d_sum = marshalDevicePtrs adata sum
in do
blk <- allocateArray (Z :. numIntervals) :: CIO (Vector e)
-- Phase 1: Split the array over multiple thread blocks and calculate
-- the final scan result from each interval.
--
when (numIntervals > 1) $ do
execute upsweep1 gamma aenv numElements blk stream
execute upsweep2 gamma aenv numIntervals (blk, blk, d_sum) stream
-- Phase 2: Re-scan the input using the carry-in value from each
-- interval sum calculated in phase 1.
--
execute kernel gamma aenv numElements (numElements, d_body, blk, d_sum) stream
| otherwise
= $internalError "scanOp" "missing multi-block kernel module(s)"
-- Forward permutation
--
permuteOp :: forall sh sh' e. (Shape sh, Shape sh', Elt e) => sh -> Array sh' e -> CIO (Array sh' e)
permuteOp !sh !dfs = do
let sh' = shape dfs
n' = size sh'
out <- allocateArray sh'
Array _ locks <- allocateArray sh' :: CIO (Array sh' Int32)
((), d_locks) <- devicePtrsOfArrayData locks :: CIO ((), CUDA.DevicePtr Int32)
liftIO $ CUDA.memsetAsync d_locks n' 0 (Just stream) -- TLM: overlap these two operations?
copyArrayAsync dfs out (Just stream)
execute kernel gamma aenv (size sh) (out, d_locks) stream
return out
-- Stencil operations. NOTE: the arguments to 'namesOfArray' must be the
-- same as those given in the function 'mkStencil[2]'.
--
stencilOp :: forall sh a b. (Shape sh, Elt a, Elt b) => Array sh a -> CIO (Array sh b)
stencilOp !arr = do
let sh = shape arr
out <- allocateArray sh
dev <- asks deviceProperties
if computeCapability dev < Compute 2 0
then marshalAccTex (namesOfArray "Stencil" (undefined :: a)) kernel arr >>
execute kernel gamma aenv (size sh) (out, sh) stream
else execute kernel gamma aenv (size sh) (out, arr) stream
--
return out
stencil2Op :: forall sh a b c. (Shape sh, Elt a, Elt b, Elt c)
=> Array sh a -> Array sh b -> CIO (Array sh c)
stencil2Op !arr1 !arr2
| Cons _ spec _ <- more
= let sh1 = shape arr1
sh2 = shape arr2
(sh, op)
| fromElt sh1 == fromElt sh2 = (sh1, spec)
| otherwise = (sh1 `intersect` sh2, kernel)
in do
out <- allocateArray sh
dev <- asks deviceProperties
if computeCapability dev < Compute 2 0
then marshalAccTex (namesOfArray "Stencil1" (undefined :: a)) op arr1 >>
marshalAccTex (namesOfArray "Stencil2" (undefined :: b)) op arr2 >>
execute op gamma aenv (size sh) (out, sh1, sh2) stream
else execute op gamma aenv (size sh) (out, arr1, arr2) stream
--
return out
| otherwise
= $internalError "stencil2Op" "missing stencil specialisation kernel"
-- Scalar expression evaluation
-- ----------------------------
executeExp :: ExecExp aenv t -> Aval aenv -> Stream -> CIO t
executeExp !exp !aenv !stream = executeOpenExp exp Empty aenv stream
executeOpenExp :: forall env aenv exp. ExecOpenExp env aenv exp -> Val env -> Aval aenv -> Stream -> CIO exp
executeOpenExp !rootExp !env !aenv !stream = travE rootExp
where
travE :: ExecOpenExp env aenv t -> CIO t
travE exp = case exp of
Var ix -> return (prj ix env)
Let bnd body -> travE bnd >>= \x -> executeOpenExp body (env `Push` x) aenv stream
Const c -> return (toElt c)
PrimConst c -> return (evalPrimConst c)
PrimApp f x -> evalPrim f <$> travE x
Tuple t -> toTuple <$> travT t
Prj ix e -> evalPrj ix . fromTuple <$> travE e
Cond p t e -> travE p >>= \x -> if x then travE t else travE e
While p f x -> while p f =<< travE x
IndexAny -> return Any
IndexNil -> return Z
IndexCons sh sz -> (:.) <$> travE sh <*> travE sz
IndexHead sh -> (\(_ :. ix) -> ix) <$> travE sh
IndexTail sh -> (\(ix :. _) -> ix) <$> travE sh
IndexSlice ix slix sh -> indexSlice ix <$> travE slix <*> travE sh
IndexFull ix slix sl -> indexFull ix <$> travE slix <*> travE sl
ToIndex sh ix -> toIndex <$> travE sh <*> travE ix
FromIndex sh ix -> fromIndex <$> travE sh <*> travE ix
Intersect sh1 sh2 -> intersect <$> travE sh1 <*> travE sh2
ShapeSize sh -> size <$> travE sh
Shape acc -> shape <$> travA acc
Index acc ix -> join $ index <$> travA acc <*> travE ix
LinearIndex acc ix -> join $ indexArray <$> travA acc <*> travE ix
Foreign _ f x -> foreign f x
-- Helpers
-- -------
travT :: Tuple (ExecOpenExp env aenv) t -> CIO t
travT tup = case tup of
NilTup -> return ()
SnocTup !t !e -> (,) <$> travT t <*> travE e
travA :: ExecOpenAcc aenv a -> CIO a
travA !acc = executeOpenAcc acc aenv stream
foreign :: ExecFun () (a -> b) -> ExecOpenExp env aenv a -> CIO b
foreign (Lam (Body f)) x = travE x >>= \e -> executeOpenExp f (Empty `Push` e) Aempty stream
foreign _ _ = error "I bless the rains down in Africa"
travF1 :: ExecOpenFun env aenv (a -> b) -> a -> CIO b
travF1 (Lam (Body f)) x = executeOpenExp f (env `Push` x) aenv stream
travF1 _ _ = error "Gonna take some time to do the things we never have"
while :: ExecOpenFun env aenv (a -> Bool) -> ExecOpenFun env aenv (a -> a) -> a -> CIO a
while !p !f !x = do
ok <- travF1 p x
if ok then while p f =<< travF1 f x
else return x
indexSlice :: (Elt slix, Elt sh, Elt sl)
=> SliceIndex (EltRepr slix) (EltRepr sl) co (EltRepr sh)
-> slix
-> sh
-> sl
indexSlice !ix !slix !sh = toElt $! restrict ix (fromElt slix) (fromElt sh)
where
restrict :: SliceIndex slix sl co sh -> slix -> sh -> sl
restrict SliceNil () () = ()
restrict (SliceAll sliceIdx) (slx, ()) (sl, sz) = (restrict sliceIdx slx sl, sz)
restrict (SliceFixed sliceIdx) (slx, _) (sl, _) = restrict sliceIdx slx sl
indexFull :: (Elt slix, Elt sh, Elt sl)
=> SliceIndex (EltRepr slix) (EltRepr sl) co (EltRepr sh)
-> slix
-> sl
-> sh
indexFull !ix !slix !sl = toElt $! extend ix (fromElt slix) (fromElt sl)
where
extend :: SliceIndex slix sl co sh -> slix -> sl -> sh
extend SliceNil () () = ()
extend (SliceAll sliceIdx) (slx, ()) (sh, sz) = (extend sliceIdx slx sh, sz)
extend (SliceFixed sliceIdx) (slx, sz) sh = (extend sliceIdx slx sh, sz)
index :: (Shape sh, Elt e) => Array sh e -> sh -> CIO e
index !arr !ix = indexArray arr (toIndex (shape arr) ix)
-- Marshalling data
-- ----------------
-- Data which can be marshalled as function arguments to a kernel invocation.
--
class Marshalable a where
marshal :: a -> CIO [CUDA.FunParam]
instance Marshalable () where
marshal () = return []
instance Marshalable CUDA.FunParam where
marshal !x = return [x]
instance ArrayElt e => Marshalable (ArrayData e) where
marshal !ad = marshalArrayData ad
instance Shape sh => Marshalable sh where
marshal !sh = marshal (reverse (shapeToList sh))
instance Marshalable a => Marshalable [a] where
marshal = concatMapM marshal
instance (Marshalable sh, Elt e) => Marshalable (Array sh e) where
marshal !(Array sh ad) = (++) <$> marshal (toElt sh :: sh) <*> marshal ad
instance (Marshalable a, Marshalable b) => Marshalable (a, b) where
marshal (!a, !b) = (++) <$> marshal a <*> marshal b
instance (Marshalable a, Marshalable b, Marshalable c) => Marshalable (a, b, c) where
marshal (!a, !b, !c)
= concat <$> sequence [marshal a, marshal b, marshal c]
instance (Marshalable a, Marshalable b, Marshalable c, Marshalable d)
=> Marshalable (a, b, c, d) where
marshal (!a, !b, !c, !d)
= concat <$> sequence [marshal a, marshal b, marshal c, marshal d]
#define primMarshalable(ty) \
instance Marshalable (ty) where { \
marshal !x = return [CUDA.VArg x] }
primMarshalable(Int)
primMarshalable(Int8)
primMarshalable(Int16)
primMarshalable(Int32)
primMarshalable(Int64)
primMarshalable(Word)
primMarshalable(Word8)
primMarshalable(Word16)
primMarshalable(Word32)
primMarshalable(Word64)
primMarshalable(Float)
primMarshalable(Double)
primMarshalable(CUDA.DevicePtr a)
-- Note [Array references in scalar code]
-- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
--
-- All CUDA devices have between 6-8KB of read-only texture memory per
-- multiprocessor. Since all arrays in Accelerate are immutable, we can always
-- access input arrays through the texture cache to reduce global memory demand
-- when accesses do not follow the regular patterns required for coalescing.
--
-- This is great for older 1.x series devices, but newer devices have a
-- dedicated L2 cache (device dependent, 256KB-1.5MB), as well as a configurable
-- L1 cache combined with shared memory (16-48KB).
--
-- For older 1.x series devices, we pass free array variables as texture
-- references, but for new devices we pass them as standard array arguments so
-- as to use the larger available caches.
--
marshalAccEnvTex :: AccKernel a -> Aval aenv -> Gamma aenv -> Stream -> CIO [CUDA.FunParam]
marshalAccEnvTex !kernel !aenv (Gamma !gamma) !stream
= flip concatMapM (Map.toList gamma)
$ \(Idx_ !(idx :: Idx aenv (Array sh e)), i) ->
do arr <- after stream (aprj idx aenv)
marshalAccTex (namesOfArray (groupOfInt i) (undefined :: e)) kernel arr
marshal (shape arr)
marshalAccTex :: (Name,[Name]) -> AccKernel a -> Array sh e -> CIO ()
marshalAccTex (_, !arrIn) (AccKernel _ _ !mdl _ _ _ _) (Array !sh !adata)
= marshalTextureData adata (R.size sh) =<< liftIO (sequence' $ map (CUDA.getTex mdl) (reverse arrIn))
marshalAccEnvArg :: Aval aenv -> Gamma aenv -> Stream -> CIO [CUDA.FunParam]
marshalAccEnvArg !aenv (Gamma !gamma) !stream
= concatMapM (\(Idx_ !idx) -> marshal =<< after stream (aprj idx aenv)) (Map.keys gamma)
-- A lazier version of 'Control.Monad.sequence'
--
sequence' :: [IO a] -> IO [a]
sequence' = foldr k (return [])
where k m ms = do { x <- m; xs <- unsafeInterleaveIO ms; return (x:xs) }
-- Generalise concatMap for teh monadz
--
concatMapM :: Monad m => (a -> m [b]) -> [a] -> m [b]
concatMapM f xs = concat `liftM` mapM f xs
-- Kernel execution
-- ----------------
-- What launch parameters should we use to execute the kernel with a number of
-- array elements?
--
configure :: AccKernel a -> Int -> (Int, Int, Int)
configure (AccKernel _ _ _ _ !cta !smem !grid) !n = (cta, grid n, smem)
-- Marshal the kernel arguments. For older 1.x devices this binds free arrays to
-- texture references, and for newer devices adds the parameters to the front of
-- the argument list
--
arguments :: Marshalable args
=> AccKernel a
-> Aval aenv
-> Gamma aenv
-> args
-> Stream
-> CIO [CUDA.FunParam]
arguments !kernel !aenv !gamma !a !stream = do
dev <- asks deviceProperties
let marshaller | computeCapability dev < Compute 2 0 = marshalAccEnvTex kernel
| otherwise = marshalAccEnvArg
--
(++) <$> marshaller aenv gamma stream <*> marshal a
-- Link the binary object implementing the computation, configure the kernel
-- launch parameters, and initiate the computation. This also handles lifting
-- and binding of array references from scalar expressions.
--
execute :: Marshalable args
=> AccKernel a -- The binary module implementing this kernel
-> Gamma aenv -- variables of arrays embedded in scalar expressions
-> Aval aenv -- the environment
-> Int -- a "size" parameter, typically number of elements in the output
-> args -- arguments to marshal to the kernel function
-> Stream -- Compute stream to execute in
-> CIO ()
execute !kernel !gamma !aenv !n !a !stream = do
args <- arguments kernel aenv gamma a stream
launch kernel (configure kernel n) args stream
-- Execute a device function, with the given thread configuration and function
-- parameters. The tuple contains (threads per block, grid size, shared memory)
--
launch :: AccKernel a -> (Int,Int,Int) -> [CUDA.FunParam] -> Stream -> CIO ()
launch (AccKernel entry !fn _ _ _ _ _) !(cta, grid, smem) !args !stream
= D.timed D.dump_exec msg (Just stream)
$ liftIO $ CUDA.launchKernel fn (grid,1,1) (cta,1,1) smem (Just stream) args
where
msg gpuTime cpuTime
= "exec: " ++ entry ++ "<<< " ++ shows grid ", " ++ shows cta ", " ++ shows smem " >>> "
++ D.elapsed gpuTime cpuTime
| kumasento/accelerate-cuda | Data/Array/Accelerate/CUDA/Execute.hs | bsd-3-clause | 28,128 | 0 | 18 | 8,699 | 8,273 | 4,134 | 4,139 | 419 | 38 |
-- | Bot that spits out core.
module Corebot where
| chrisdone/corebot | src/Corebot.hs | bsd-3-clause | 52 | 0 | 2 | 11 | 5 | 4 | 1 | 1 | 0 |
{-# LANGUAGE KindSignatures #-}
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE FunctionalDependencies #-}
{-# LANGUAGE ScopedTypeVariables #-}
{-# LANGUAGE DeriveFunctor #-}
{-# LANGUAGE FlexibleContexts #-}
{-# LANGUAGE TypeSynonymInstances #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE GeneralizedNewtypeDeriving #-}
{-# LANGUAGE OverloadedStrings #-}
-- | A demo of annotations
module Text.PrettyPrint.Final.Demos.ListDemo () where
import Control.Monad
import Control.Applicative
import Control.Monad.Identity
import Control.Monad.Reader
import Control.Monad.Writer
import Control.Monad.State
import Control.Monad.RWS
import Data.List
import Data.String (IsString(..))
import Data.Text (Text)
import qualified Data.Text as T
import System.Console.ANSI
import Text.PrettyPrint.Final
import Text.PrettyPrint.Final.Rendering.Console
-- Constructor names or built-in syntax
data HsAnn = Ctor | Stx
deriving (Eq, Ord, Show)
env0 :: Monoid fmt => PEnv Int a fmt
env0 = PEnv
{ maxWidth = 80
, maxRibbon = 60
, layout = Break
, failure = CantFail
, nesting = 0
, formatting = mempty
, formatAnn = const mempty
}
state0 :: PState Int ()
state0 = PState
{ curLine = []
}
-- For plain text pretty printing
newtype DocM a = DocM { unDocM :: RWST (PEnv Int HsAnn ()) (POut Int HsAnn) (PState Int ()) Maybe a }
deriving
( Functor, Applicative, Monad
, MonadReader (PEnv Int HsAnn ()), MonadWriter (POut Int HsAnn), MonadState (PState Int ()), Alternative
)
instance MonadPretty Int HsAnn () DocM
instance IsString (DocM ()) where
fromString = text . fromString
runDocM :: PEnv Int HsAnn () -> PState Int () -> DocM a -> Maybe (PState Int (), POut Int HsAnn, a)
runDocM e s d = (\(a,s',o) -> (s',o,a)) <$> runRWST (unDocM d) e s
execDoc :: Doc -> POut Int HsAnn
execDoc d =
let rM = runDocM env0 state0 d
in case rM of
Nothing -> PAtom $ AChunk $ CText "<internal pretty printing error>"
Just (_, o, ()) -> o
type Doc = DocM ()
instance Monoid Doc where
mempty = return ()
mappend = (>>)
class Pretty a where
pretty :: a -> Doc
instance Pretty Doc where
pretty = id
instance Measure Int () DocM where
measure = return . runIdentity . measure
instance Pretty Text where
pretty = annotate Ctor . text . T.pack . show
instance (Pretty a) => Pretty [a] where
pretty = collection (annotate Stx "[") (annotate Stx "]") (annotate Stx ",") . map pretty
toSGR :: HsAnn -> [SGR]
toSGR Ctor = [SetConsoleIntensity BoldIntensity, SetColor Foreground Vivid Red]
toSGR Stx = [SetConsoleIntensity BoldIntensity, SetColor Foreground Vivid Black]
updateColor :: forall ann . StateT [HsAnn] IO ()
updateColor =
lift . setSGR =<< mconcat . map toSGR . reverse <$> get
openTag :: HsAnn -> StateT [HsAnn] IO ()
openTag ann = modify (ann:) >> updateColor
closeTag :: HsAnn -> StateT [HsAnn] IO ()
closeTag _ = modify tail >> updateColor
renderAnnotation :: HsAnn -> StateT [HsAnn] IO () -> StateT [HsAnn] IO ()
renderAnnotation a o = openTag a >> o >> closeTag a
dumpList :: Doc -> IO ()
dumpList = dumpDoc toSGR renderAnnotation . execDoc
---------------
-- Test docs --
---------------
shortList :: [[Text]]
shortList = [["a", "b", "c"], [], ["longer"]]
longList :: [[Text]]
longList = [map (T.pack . show) [1..10], [], map (T.pack . flip replicate 'a') [1..10]]
-- To try, eval dumpDoc (pretty shortList) or dumpDoc (pretty longList) in console GHCI
| david-christiansen/final-pretty-printer | Text/PrettyPrint/Final/Demos/ListDemo.hs | mit | 3,428 | 0 | 12 | 633 | 1,178 | 646 | 532 | 86 | 2 |
predecessor = predecessor
successor = successor
| evolutics/haskell-formatter | testsuite/resources/source/comments/empty_lines/between_top_level_functions/without_type_signature/0/Output.hs | gpl-3.0 | 48 | 0 | 4 | 6 | 11 | 6 | 5 | 2 | 1 |
{-# OPTIONS -fno-warn-overlapping-patterns #-}
module Test.TestUtil
(
module Test.TestUtil,
module Test.Framework,
module Test.Framework.Providers.QuickCheck2
) where
import Test.Framework
import Test.Framework.Providers.API
import Test.Framework.Providers.QuickCheck2
data Result = Pass | Fail String
instance Show Result where
show Pass = "passed"
show (Fail s) = "failed: " ++ s
instance TestResultlike () Result where
testSucceeded Pass = True
testSucceeded (Fail _) = False
instance Testlike () Result (IO Result) where
testTypeName _ = "Cases"
runTest _ ior = do
r <- ior
return (Finished r,return ())
ioTest :: String -> IO Result -> Test
ioTest = Test
pureTest :: String -> Result -> Test
pureTest name result = ioTest name (return result)
diff :: (Show a,Eq a) => a -> a -> Result
diff expected found | expected == found = Pass
diff expected found = Fail ("expected " ++ (show expected) ++ " but found " ++ (show found))
| jwiegley/ghc-release | libraries/time/test/TestUtil.hs | gpl-3.0 | 999 | 0 | 11 | 214 | 336 | 177 | 159 | -1 | -1 |
{-# LANGUAGE OverloadedStrings #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- Module : Test.AWS.CloudFront.Internal
-- Copyright : (c) 2013-2015 Brendan Hay
-- License : This Source Code Form is subject to the terms of
-- the Mozilla Public License, v. 2.0.
-- A copy of the MPL can be found in the LICENSE file or
-- you can obtain it at http://mozilla.org/MPL/2.0/.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : experimental
-- Portability : non-portable (GHC extensions)
module Test.AWS.CloudFront.Internal where
import Test.AWS.Prelude
| fmapfmapfmap/amazonka | amazonka-cloudfront/test/Test/AWS/CloudFront/Internal.hs | mpl-2.0 | 627 | 0 | 4 | 140 | 25 | 21 | 4 | 4 | 0 |
{-# LANGUAGE NoMonomorphismRestriction, FlexibleContexts, DeriveDataTypeable #-}
-- | Diagrams-AST provides a data oriented interface to the <http://hackage.haskell.org/package/diagrams> package.
module Diagrams.AST (
-- Functions
outputImage,
runImage,
getAngleFraction,
getAngleRadians,
getAngleDegrees,
-- Data Types
Image (..),
Modifier (..),
Images (..),
Path (..),
Shape (..),
ColorData (..),
Fill (..),
Alignment (..),
Angle (..)
)
where
-- Diagram Imports
import qualified Diagrams.Prelude as D
import qualified Diagrams.Path as P
import qualified Diagrams.TwoD.Path as P2
import qualified Diagrams.TwoD.Arc as A
import qualified Diagrams.TwoD.Align as L
import qualified Diagrams.Backend.Cairo as C
import qualified Diagrams.TwoD.Text as T
import qualified Graphics.Rendering.Diagrams.Points as P3
import Diagrams.Prelude ((|||), (===))
-- Data Imports
import Data.Monoid
import Data.List (foldl')
-- Meta
import Data.Generics.Uniplate.Data
import Data.Data
--- Data Types
data Image = Blank
| Shape Shape
| Modifier Modifier Image
| Images Images deriving (Show, Eq, Ord, Data, Typeable)
data Modifier = Foreground ColorData
| LineColor ColorData
| LineWidth Double
| Dashing [Double] Double
| Translate Double Double
| Scale Double Double
| Rotate Angle
| Pad Double
| Freeze
| Origin
| Align Alignment
| Changes [Modifier] deriving (Show, Eq, Ord, Data, Typeable)
data Images = Atop Image Image
| NextTo Image Image
| Above Image Image
| Layers [Image]
| Horizontal [Image]
| Vertical [Image] deriving (Show, Eq, Ord, Data, Typeable)
data Shape = Text String | Circle | Square | Path Fill Path deriving (Show, Eq, Ord, Data, Typeable)
data Path = Offsets [(Double,Double)]
| Points [(Double,Double)]
| Arc Angle Angle deriving (Show, Eq, Ord, Data, Typeable)
data ColorData = RGBA Double Double Double Double -- ^ Red, Green, Blue, Alpha
| RAA Double Angle Angle Double -- ^ Radius, Blue\/Green, (Blue\/Green)\/Red
deriving (Show, Eq, Ord, Data, Typeable)
data Fill = Closed | Open deriving (Show, Eq, Ord, Data, Typeable)
-- | Alignment of the origin of an 'Image'.
data Alignment = L -- ^ Left
| R -- ^ Right
| T -- ^ Top
| B -- ^ Bottom
| TL -- ^ Top-Left
| TR -- ^ Top-Right
| BL -- ^ Bottom-Left
| BR -- ^ Bottom-Right
| C -- ^ Center
| CX -- ^ X-Centered
| CY -- ^ Y-Centered
| X Double -- ^ X-Proportion (Fraction -1 to 1)
| Y Double -- ^ Y-Proportion (Fraction -1 to 1)
deriving (Show, Eq, Ord, Data, Typeable)
-- | Angles are instances of Num. 'fromInteger' interprets its argument as a fraction of a full circle.
data Angle = Fraction Double | Radians Double | Degrees Double deriving (Show, Eq, Ord, Data, Typeable)
instance Num Angle
where
fromInteger x = Fraction $ fromInteger x
x + y = Fraction $ (getAngleFraction x) + (getAngleFraction y)
x * y = Fraction $ (getAngleFraction x) * (getAngleFraction y)
abs x = Fraction $ abs (getAngleFraction x)
signum x
| x' > 0 = 1
| x' < 0 = -1
| otherwise = 0
where x' = getAngleFraction x
instance D.Angle Angle where
toCircleFrac a = D.CircleFrac $ getAngleFraction a
fromCircleFrac (D.CircleFrac d) = Fraction d
-- | 'getAngleFraction' returns the fraction of a full circle for any angle.
getAngleFraction :: Angle -> Double
getAngleFraction (Fraction x) = x
getAngleFraction (Radians x) = x / (2*pi)
getAngleFraction (Degrees x) = x / 360
getAngleRadians = (* 2) . (* pi) . getAngleFraction
getAngleDegrees = (* 360) . getAngleFraction
--- Instances
instance D.Color ColorData where
colorToRGBA (RGBA r g b a) = (r, g, b, a)
colorToRGBA (RAA r g e a) = ( r * cos g' * cos e',
r * cos g' * sin e',
r * sin g',
a ) where [g', e'] = map getAngleRadians [g, e]
---- Run ADT Functions
-- | 'outputImage' renders a PNG to the file supplied.
outputImage :: String -> Int -> Int -> Image -> IO ()
outputImage name width height image = do
-- Is a Result type in Cairo a pair a la State?
-- No idea why I should need to call fst otherwise
-- D.renderDia :: b -> Options b v -> AnnDiagram b v m -> Result b v
fst $ D.renderDia C.Cairo (C.CairoOptions name (C.PNG (width, height))) (runImage image)
--- Main runner
{-
runImage :: (D.Renderable Diagrams.TwoD.Ellipse.Ellipse b, D.Renderable (P.Path D.R2) b, D.Backend b D.R2) =>
Image -> D.Diagram b D.R2
-}
runImage (Shape s) = runShape s
runImage (Modifier m i) = runModifier m (runImage i)
runImage (Images c) = runCombiner c
runImage Blank = mempty
--- Internal runners
runCombiner (Atop l r) = runImage l `D.atop` runImage r
runCombiner (NextTo l r) = runImage l ||| runImage r
runCombiner (Above t b) = runImage t === runImage b
runCombiner (Layers l) = mconcat . map runImage $ l
runCombiner (Horizontal l) = D.hcat (map runImage l)
runCombiner (Vertical l) = D.vcat (map runImage l)
runShape (Text s) = T.text s
runShape Circle = D.circle 1
runShape Square = D.square 1
runShape (Path Closed p) = P2.stroke $ P.close $ runPath p
runShape (Path Open p) = P2.stroke $ P.open $ runPath p
runModifier (Foreground c) = D.fillColor c
runModifier (LineColor c) = D.lineColor c
runModifier (LineWidth w) = D.lw w
runModifier (Dashing a w) = D.dashing a w
runModifier (Translate x y) = D.translate (x, y)
runModifier (Rotate a) = D.rotateBy (D.CircleFrac $ getAngleFraction a)
runModifier (Scale x y) = D.scaleX x . D.scaleY y
runModifier (Pad r) = D.pad r
runModifier (Align a) = runAlign a
runModifier (Changes l) = foldl' (flip (.)) id . map runModifier $ l
runModifier Origin = D.showOrigin
runModifier Freeze = D.freeze
runPath (Offsets l) = P.fromOffsets l
runPath (Points l) = (P.fromVertices . map P3.P) l
runPath (Arc b e) = A.arc b e
runAlign L = L.alignL
runAlign R = L.alignR
runAlign T = L.alignT
runAlign B = L.alignB
runAlign TL = L.alignTL
runAlign TR = L.alignTR
runAlign BL = L.alignBL
runAlign BR = L.alignBR
runAlign C = L.centerXY
runAlign CX = L.centerX
runAlign CY = L.centerY
runAlign (X x) = L.alignX x
runAlign (Y y) = L.alignY y
| beni55/Diagrams-AST | src/Diagrams/AST.hs | bsd-3-clause | 7,010 | 0 | 13 | 2,143 | 2,097 | 1,133 | 964 | 145 | 1 |
-- Copyright (c) 2015 Eric McCorkle. All rights reserved.
--
-- Redistribution and use in source and binary forms, with or without
-- modification, are permitted provided that the following conditions
-- are met:
--
-- 1. Redistributions of source code must retain the above copyright
-- notice, this list of conditions and the following disclaimer.
--
-- 2. Redistributions in binary form must reproduce the above copyright
-- notice, this list of conditions and the following disclaimer in the
-- documentation and/or other materials provided with the distribution.
--
-- 3. Neither the name of the author nor the names of any contributors
-- may be used to endorse or promote products derived from this software
-- without specific prior written permission.
--
-- THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS''
-- AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
-- TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
-- PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS
-- OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-- SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-- LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
-- USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-- ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
-- OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
-- OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
-- SUCH DAMAGE.
{-# OPTIONS_GHC -funbox-strict-fields -Wall -Werror #-}
{-# LANGUAGE MultiParamTypeClasses, FlexibleInstances #-}
module Control.Monad.ProofNames.Class(
MonadProofNames(..)
) where
import Control.Monad.Reader
import Control.Monad.State
import Control.Monad.Writer
import Language.Salt.Core.Syntax
-- | A monad class for names used by the proof checker.
class Monad m => MonadProofNames sym m where
-- | The term representing the implies proposition.
impliesProp :: m (Term sym sym)
-- | The symbol for the name "premise", an argument to the implies
-- function.
premiseName :: m sym
-- | The symbol for the name "consequence", an argument to the
-- implies function.
consequenceName :: m sym
instance MonadProofNames sym m => MonadProofNames sym (ReaderT s m) where
impliesProp = lift impliesProp
premiseName = lift premiseName
consequenceName = lift consequenceName
instance MonadProofNames sym m => MonadProofNames sym (StateT s m) where
impliesProp = lift impliesProp
premiseName = lift premiseName
consequenceName = lift consequenceName
instance (Monoid s, MonadProofNames sym m) =>
MonadProofNames sym (WriterT s m) where
impliesProp = lift impliesProp
premiseName = lift premiseName
consequenceName = lift consequenceName
| emc2/saltlang | src/salt/Control/Monad/ProofNames/Class.hs | bsd-3-clause | 2,847 | 0 | 9 | 497 | 290 | 170 | 120 | 25 | 0 |
{-# LANGUAGE GADTs, RecordWildCards, MagicHash, ScopedTypeVariables, CPP,
UnboxedTuples #-}
{-# OPTIONS_GHC -fno-warn-name-shadowing #-}
-- |
-- Execute GHCi messages.
--
-- For details on Remote GHCi, see Note [Remote GHCi] in
-- compiler/ghci/GHCi.hs.
--
module GHCi.Run
( run, redirectInterrupts
) where
import Prelude -- See note [Why do we import Prelude here?]
import GHCi.CreateBCO
import GHCi.InfoTable
import GHCi.FFI
import GHCi.Message
import GHCi.ObjLink
import GHCi.RemoteTypes
import GHCi.TH
import GHCi.BreakArray
import GHCi.StaticPtrTable
import Control.Concurrent
import Control.DeepSeq
import Control.Exception
import Control.Monad
import Data.Binary
import Data.Binary.Get
import Data.ByteString (ByteString)
import qualified Data.ByteString.Unsafe as B
import GHC.Exts
import GHC.Exts.Heap
import GHC.Stack
import Foreign hiding (void)
import Foreign.C
import GHC.Conc.Sync
import GHC.IO hiding ( bracket )
import System.Mem.Weak ( deRefWeak )
import Unsafe.Coerce
-- -----------------------------------------------------------------------------
-- Implement messages
foreign import ccall "revertCAFs" rts_revertCAFs :: IO ()
-- Make it "safe", just in case
run :: Message a -> IO a
run m = case m of
InitLinker -> initObjLinker RetainCAFs
RtsRevertCAFs -> rts_revertCAFs
LookupSymbol str -> fmap toRemotePtr <$> lookupSymbol str
LookupClosure str -> lookupClosure str
LoadDLL str -> loadDLL str
LoadArchive str -> loadArchive str
LoadObj str -> loadObj str
UnloadObj str -> unloadObj str
AddLibrarySearchPath str -> toRemotePtr <$> addLibrarySearchPath str
RemoveLibrarySearchPath ptr -> removeLibrarySearchPath (fromRemotePtr ptr)
ResolveObjs -> resolveObjs
FindSystemLibrary str -> findSystemLibrary str
CreateBCOs bcos -> createBCOs (concatMap (runGet get) bcos)
FreeHValueRefs rs -> mapM_ freeRemoteRef rs
AddSptEntry fpr r -> localRef r >>= sptAddEntry fpr
EvalStmt opts r -> evalStmt opts r
ResumeStmt opts r -> resumeStmt opts r
AbandonStmt r -> abandonStmt r
EvalString r -> evalString r
EvalStringToString r s -> evalStringToString r s
EvalIO r -> evalIO r
MkCostCentres mod ccs -> mkCostCentres mod ccs
CostCentreStackInfo ptr -> ccsToStrings (fromRemotePtr ptr)
NewBreakArray sz -> mkRemoteRef =<< newBreakArray sz
EnableBreakpoint ref ix b -> do
arr <- localRef ref
_ <- if b then setBreakOn arr ix else setBreakOff arr ix
return ()
BreakpointStatus ref ix -> do
arr <- localRef ref; r <- getBreak arr ix
case r of
Nothing -> return False
Just w -> return (w /= 0)
GetBreakpointVar ref ix -> do
aps <- localRef ref
mapM mkRemoteRef =<< getIdValFromApStack aps ix
MallocData bs -> mkString bs
MallocStrings bss -> mapM mkString0 bss
PrepFFI conv args res -> toRemotePtr <$> prepForeignCall conv args res
FreeFFI p -> freeForeignCallInfo (fromRemotePtr p)
MkConInfoTable ptrs nptrs tag ptrtag desc ->
toRemotePtr <$> mkConInfoTable ptrs nptrs tag ptrtag desc
StartTH -> startTH
GetClosure ref -> do
clos <- getClosureData =<< localRef ref
mapM (\(Box x) -> mkRemoteRef (HValue x)) clos
Seq ref -> tryEval (void $ evaluate =<< localRef ref)
_other -> error "GHCi.Run.run"
evalStmt :: EvalOpts -> EvalExpr HValueRef -> IO (EvalStatus [HValueRef])
evalStmt opts expr = do
io <- mkIO expr
sandboxIO opts $ do
rs <- unsafeCoerce io :: IO [HValue]
mapM mkRemoteRef rs
where
mkIO (EvalThis href) = localRef href
mkIO (EvalApp l r) = do
l' <- mkIO l
r' <- mkIO r
return ((unsafeCoerce l' :: HValue -> HValue) r')
evalIO :: HValueRef -> IO (EvalResult ())
evalIO r = do
io <- localRef r
tryEval (unsafeCoerce io :: IO ())
evalString :: HValueRef -> IO (EvalResult String)
evalString r = do
io <- localRef r
tryEval $ do
r <- unsafeCoerce io :: IO String
evaluate (force r)
evalStringToString :: HValueRef -> String -> IO (EvalResult String)
evalStringToString r str = do
io <- localRef r
tryEval $ do
r <- (unsafeCoerce io :: String -> IO String) str
evaluate (force r)
-- When running a computation, we redirect ^C exceptions to the running
-- thread. ToDo: we might want a way to continue even if the target
-- thread doesn't die when it receives the exception... "this thread
-- is not responding".
--
-- Careful here: there may be ^C exceptions flying around, so we start the new
-- thread blocked (forkIO inherits mask from the parent, #1048), and unblock
-- only while we execute the user's code. We can't afford to lose the final
-- putMVar, otherwise deadlock ensues. (#1583, #1922, #1946)
sandboxIO :: EvalOpts -> IO a -> IO (EvalStatus a)
sandboxIO opts io = do
-- We are running in uninterruptibleMask
breakMVar <- newEmptyMVar
statusMVar <- newEmptyMVar
withBreakAction opts breakMVar statusMVar $ do
let runIt = measureAlloc $ tryEval $ rethrow opts $ clearCCS io
if useSandboxThread opts
then do
tid <- forkIO $ do unsafeUnmask runIt >>= putMVar statusMVar
-- empty: can't block
redirectInterrupts tid $ unsafeUnmask $ takeMVar statusMVar
else
-- GLUT on OS X needs to run on the main thread. If you
-- try to use it from another thread then you just get a
-- white rectangle rendered. For this, or anything else
-- with such restrictions, you can turn the GHCi sandbox off
-- and things will be run in the main thread.
--
-- BUT, note that the debugging features (breakpoints,
-- tracing, etc.) need the expression to be running in a
-- separate thread, so debugging is only enabled when
-- using the sandbox.
runIt
-- We want to turn ^C into a break when -fbreak-on-exception is on,
-- but it's an async exception and we only break for sync exceptions.
-- Idea: if we catch and re-throw it, then the re-throw will trigger
-- a break. Great - but we don't want to re-throw all exceptions, because
-- then we'll get a double break for ordinary sync exceptions (you'd have
-- to :continue twice, which looks strange). So if the exception is
-- not "Interrupted", we unset the exception flag before throwing.
--
rethrow :: EvalOpts -> IO a -> IO a
rethrow EvalOpts{..} io =
catch io $ \se -> do
-- If -fbreak-on-error, we break unconditionally,
-- but with care of not breaking twice
if breakOnError && not breakOnException
then poke exceptionFlag 1
else case fromException se of
-- If it is a "UserInterrupt" exception, we allow
-- a possible break by way of -fbreak-on-exception
Just UserInterrupt -> return ()
-- In any other case, we don't want to break
_ -> poke exceptionFlag 0
throwIO se
--
-- While we're waiting for the sandbox thread to return a result, if
-- the current thread receives an asynchronous exception we re-throw
-- it at the sandbox thread and continue to wait.
--
-- This is for two reasons:
--
-- * So that ^C interrupts runStmt (e.g. in GHCi), allowing the
-- computation to run its exception handlers before returning the
-- exception result to the caller of runStmt.
--
-- * clients of the GHC API can terminate a runStmt in progress
-- without knowing the ThreadId of the sandbox thread (#1381)
--
-- NB. use a weak pointer to the thread, so that the thread can still
-- be considered deadlocked by the RTS and sent a BlockedIndefinitely
-- exception. A symptom of getting this wrong is that conc033(ghci)
-- will hang.
--
redirectInterrupts :: ThreadId -> IO a -> IO a
redirectInterrupts target wait = do
wtid <- mkWeakThreadId target
wait `catch` \e -> do
m <- deRefWeak wtid
case m of
Nothing -> wait
Just target -> do throwTo target (e :: SomeException); wait
measureAlloc :: IO (EvalResult a) -> IO (EvalStatus a)
measureAlloc io = do
setAllocationCounter maxBound
a <- io
ctr <- getAllocationCounter
let allocs = fromIntegral (maxBound::Int64) - fromIntegral ctr
return (EvalComplete allocs a)
-- Exceptions can't be marshaled because they're dynamically typed, so
-- everything becomes a String.
tryEval :: IO a -> IO (EvalResult a)
tryEval io = do
e <- try io
case e of
Left ex -> return (EvalException (toSerializableException ex))
Right a -> return (EvalSuccess a)
-- This function sets up the interpreter for catching breakpoints, and
-- resets everything when the computation has stopped running. This
-- is a not-very-good way to ensure that only the interactive
-- evaluation should generate breakpoints.
withBreakAction :: EvalOpts -> MVar () -> MVar (EvalStatus b) -> IO a -> IO a
withBreakAction opts breakMVar statusMVar act
= bracket setBreakAction resetBreakAction (\_ -> act)
where
setBreakAction = do
stablePtr <- newStablePtr onBreak
poke breakPointIOAction stablePtr
when (breakOnException opts) $ poke exceptionFlag 1
when (singleStep opts) $ setStepFlag
return stablePtr
-- Breaking on exceptions is not enabled by default, since it
-- might be a bit surprising. The exception flag is turned off
-- as soon as it is hit, or in resetBreakAction below.
onBreak :: BreakpointCallback
onBreak ix# uniq# is_exception apStack = do
tid <- myThreadId
let resume = ResumeContext
{ resumeBreakMVar = breakMVar
, resumeStatusMVar = statusMVar
, resumeThreadId = tid }
resume_r <- mkRemoteRef resume
apStack_r <- mkRemoteRef apStack
ccs <- toRemotePtr <$> getCCSOf apStack
putMVar statusMVar $ EvalBreak is_exception apStack_r (I# ix#) (I# uniq#) resume_r ccs
takeMVar breakMVar
resetBreakAction stablePtr = do
poke breakPointIOAction noBreakStablePtr
poke exceptionFlag 0
resetStepFlag
freeStablePtr stablePtr
resumeStmt
:: EvalOpts -> RemoteRef (ResumeContext [HValueRef])
-> IO (EvalStatus [HValueRef])
resumeStmt opts hvref = do
ResumeContext{..} <- localRef hvref
withBreakAction opts resumeBreakMVar resumeStatusMVar $
mask_ $ do
putMVar resumeBreakMVar () -- this awakens the stopped thread...
redirectInterrupts resumeThreadId $ takeMVar resumeStatusMVar
-- when abandoning a computation we have to
-- (a) kill the thread with an async exception, so that the
-- computation itself is stopped, and
-- (b) fill in the MVar. This step is necessary because any
-- thunks that were under evaluation will now be updated
-- with the partial computation, which still ends in takeMVar,
-- so any attempt to evaluate one of these thunks will block
-- unless we fill in the MVar.
-- (c) wait for the thread to terminate by taking its status MVar. This
-- step is necessary to prevent race conditions with
-- -fbreak-on-exception (see #5975).
-- See test break010.
abandonStmt :: RemoteRef (ResumeContext [HValueRef]) -> IO ()
abandonStmt hvref = do
ResumeContext{..} <- localRef hvref
killThread resumeThreadId
putMVar resumeBreakMVar ()
_ <- takeMVar resumeStatusMVar
return ()
foreign import ccall "&rts_stop_next_breakpoint" stepFlag :: Ptr CInt
foreign import ccall "&rts_stop_on_exception" exceptionFlag :: Ptr CInt
setStepFlag :: IO ()
setStepFlag = poke stepFlag 1
resetStepFlag :: IO ()
resetStepFlag = poke stepFlag 0
type BreakpointCallback
= Int# -- the breakpoint index
-> Int# -- the module uniq
-> Bool -- exception?
-> HValue -- the AP_STACK, or exception
-> IO ()
foreign import ccall "&rts_breakpoint_io_action"
breakPointIOAction :: Ptr (StablePtr BreakpointCallback)
noBreakStablePtr :: StablePtr BreakpointCallback
noBreakStablePtr = unsafePerformIO $ newStablePtr noBreakAction
noBreakAction :: BreakpointCallback
noBreakAction _ _ False _ = putStrLn "*** Ignoring breakpoint"
noBreakAction _ _ True _ = return () -- exception: just continue
-- Malloc and copy the bytes. We don't have any way to monitor the
-- lifetime of this memory, so it just leaks.
mkString :: ByteString -> IO (RemotePtr ())
mkString bs = B.unsafeUseAsCStringLen bs $ \(cstr,len) -> do
ptr <- mallocBytes len
copyBytes ptr cstr len
return (castRemotePtr (toRemotePtr ptr))
mkString0 :: ByteString -> IO (RemotePtr ())
mkString0 bs = B.unsafeUseAsCStringLen bs $ \(cstr,len) -> do
ptr <- mallocBytes (len+1)
copyBytes ptr cstr len
pokeElemOff (ptr :: Ptr CChar) len 0
return (castRemotePtr (toRemotePtr ptr))
mkCostCentres :: String -> [(String,String)] -> IO [RemotePtr CostCentre]
#if defined(PROFILING)
mkCostCentres mod ccs = do
c_module <- newCString mod
mapM (mk_one c_module) ccs
where
mk_one c_module (decl_path,srcspan) = do
c_name <- newCString decl_path
c_srcspan <- newCString srcspan
toRemotePtr <$> c_mkCostCentre c_name c_module c_srcspan
foreign import ccall unsafe "mkCostCentre"
c_mkCostCentre :: Ptr CChar -> Ptr CChar -> Ptr CChar -> IO (Ptr CostCentre)
#else
mkCostCentres _ _ = return []
#endif
getIdValFromApStack :: HValue -> Int -> IO (Maybe HValue)
getIdValFromApStack apStack (I# stackDepth) = do
case getApStackVal# apStack stackDepth of
(# ok, result #) ->
case ok of
0# -> return Nothing -- AP_STACK not found
_ -> return (Just (unsafeCoerce# result))
| sdiehl/ghc | libraries/ghci/GHCi/Run.hs | bsd-3-clause | 13,446 | 0 | 18 | 2,942 | 3,055 | 1,492 | 1,563 | 232 | 38 |
{-# LANGUAGE TemplateHaskell #-}
{-# LANGUAGE GADTs #-}
{-# LANGUAGE GeneralizedNewtypeDeriving #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE PatternGuards #-}
{-# LANGUAGE CPP #-}
{-# LANGUAGE TypeFamilies #-}
{-# LANGUAGE EmptyDataDecls #-}
module Text.Css where
import Data.List (intersperse, intercalate)
import Data.Text.Lazy.Builder (Builder, singleton, toLazyText, fromLazyText, fromString)
import qualified Data.Text.Lazy as TL
import qualified Data.Text.Lazy.Builder as TLB
import Data.Monoid (Monoid, mconcat, mappend, mempty)
import Data.Text (Text)
import qualified Data.Text as T
import Language.Haskell.TH.Syntax
import System.IO.Unsafe (unsafePerformIO)
import Text.ParserCombinators.Parsec (Parser, parse)
import Text.Shakespeare.Base hiding (Scope)
import Language.Haskell.TH
import Control.Applicative ((<$>), (<*>))
import Control.Arrow ((***), second)
import Text.IndentToBrace (i2b)
import Data.Functor.Identity (runIdentity)
import Text.Shakespeare (VarType (..))
#if MIN_VERSION_base(4,5,0)
import Data.Monoid ((<>))
#else
(<>) :: Monoid m => m -> m -> m
(<>) = mappend
{-# INLINE (<>) #-}
#endif
type CssUrl url = (url -> [(T.Text, T.Text)] -> T.Text) -> Css
type DList a = [a] -> [a]
-- FIXME great use case for data kinds
data Resolved
data Unresolved
type family Selector a
type instance Selector Resolved = Builder
type instance Selector Unresolved = [Contents]
type family ChildBlocks a
type instance ChildBlocks Resolved = ()
type instance ChildBlocks Unresolved = [(HasLeadingSpace, Block Unresolved)]
type HasLeadingSpace = Bool
type family Str a
type instance Str Resolved = Builder
type instance Str Unresolved = Contents
type family Mixins a
type instance Mixins Resolved = ()
type instance Mixins Unresolved = [Deref]
data Block a = Block
{ blockSelector :: !(Selector a)
, blockAttrs :: ![Attr a]
, blockBlocks :: !(ChildBlocks a)
, blockMixins :: !(Mixins a)
}
data Mixin = Mixin
{ mixinAttrs :: ![Attr Resolved]
, mixinBlocks :: ![Block Resolved]
}
instance Monoid Mixin where
mempty = Mixin mempty mempty
mappend (Mixin a x) (Mixin b y) = Mixin (a ++ b) (x ++ y)
data TopLevel a where
TopBlock :: !(Block a) -> TopLevel a
TopAtBlock :: !String -- name e.g., media
-> !(Str a) -- selector
-> ![Block a]
-> TopLevel a
TopAtDecl :: !String -> !(Str a) -> TopLevel a
TopVar :: !String -> !String -> TopLevel Unresolved
data Attr a = Attr
{ attrKey :: !(Str a)
, attrVal :: !(Str a)
}
data Css = CssWhitespace ![TopLevel Resolved]
| CssNoWhitespace ![TopLevel Resolved]
data Content = ContentRaw String
| ContentVar Deref
| ContentUrl Deref
| ContentUrlParam Deref
| ContentMixin Deref
deriving (Show, Eq)
type Contents = [Content]
data CDData url = CDPlain Builder
| CDUrl url
| CDUrlParam (url, [(Text, Text)])
| CDMixin Mixin
pack :: String -> Text
pack = T.pack
#if !MIN_VERSION_text(0, 11, 2)
{-# NOINLINE pack #-}
#endif
fromText :: Text -> Builder
fromText = TLB.fromText
{-# NOINLINE fromText #-}
class ToCss a where
toCss :: a -> Builder
instance ToCss [Char] where toCss = fromLazyText . TL.pack
instance ToCss Text where toCss = fromText
instance ToCss TL.Text where toCss = fromLazyText
-- | Determine which identifiers are used by the given template, useful for
-- creating systems like yesod devel.
cssUsedIdentifiers :: Bool -- ^ perform the indent-to-brace conversion
-> Parser [TopLevel Unresolved]
-> String
-> [(Deref, VarType)]
cssUsedIdentifiers toi2b parseBlocks s' =
concat $ runIdentity $ mapM (getVars scope0) contents
where
s = if toi2b then i2b s' else s'
a = either (error . show) id $ parse parseBlocks s s
(scope0, contents) = go a
go :: [TopLevel Unresolved]
-> (Scope, [Content])
go [] = ([], [])
go (TopAtDecl dec cs:rest) =
(scope, rest'')
where
(scope, rest') = go rest
rest'' =
ContentRaw ('@' : dec ++ " ")
: cs
++ ContentRaw ";"
: rest'
go (TopAtBlock _ _ blocks:rest) =
(scope1 ++ scope2, rest1 ++ rest2)
where
(scope1, rest1) = go (map TopBlock blocks)
(scope2, rest2) = go rest
go (TopBlock (Block x y z mixins):rest) =
(scope1 ++ scope2, rest0 ++ rest1 ++ rest2 ++ restm)
where
rest0 = intercalate [ContentRaw ","] x ++ concatMap go' y
(scope1, rest1) = go (map (TopBlock . snd) z)
(scope2, rest2) = go rest
restm = map ContentMixin mixins
go (TopVar k v:rest) =
((k, v):scope, rest')
where
(scope, rest') = go rest
go' (Attr k v) = k ++ v
cssFileDebug :: Bool -- ^ perform the indent-to-brace conversion
-> Q Exp
-> Parser [TopLevel Unresolved]
-> FilePath
-> Q Exp
cssFileDebug toi2b parseBlocks' parseBlocks fp = do
s <- fmap TL.unpack $ qRunIO $ readUtf8File fp
#ifdef GHC_7_4
qAddDependentFile fp
#endif
let vs = cssUsedIdentifiers toi2b parseBlocks s
c <- mapM vtToExp vs
cr <- [|cssRuntime toi2b|]
parseBlocks'' <- parseBlocks'
return $ cr `AppE` parseBlocks'' `AppE` (LitE $ StringL fp) `AppE` ListE c
combineSelectors :: HasLeadingSpace
-> [Contents]
-> [Contents]
-> [Contents]
combineSelectors hsl a b = do
a' <- a
b' <- b
return $ a' ++ addSpace b'
where
addSpace
| hsl = (ContentRaw " " :)
| otherwise = id
blockRuntime :: [(Deref, CDData url)]
-> (url -> [(Text, Text)] -> Text)
-> Block Unresolved
-> Either String (DList (Block Resolved))
-- FIXME share code with blockToCss
blockRuntime cd render' (Block x attrs z mixinsDerefs) = do
mixins <- mapM getMixin mixinsDerefs
x' <- mapM go' $ intercalate [ContentRaw ","] x
attrs' <- mapM resolveAttr attrs
z' <- mapM (subGo x) z -- FIXME use difflists again
Right $ \rest -> Block
{ blockSelector = mconcat x'
, blockAttrs = concat $ attrs' : map mixinAttrs mixins
, blockBlocks = ()
, blockMixins = ()
} : foldr ($) rest z'
{-
(:) (Css' (mconcat $ map go' $ intercalate [ContentRaw "," ] x) (map go'' y))
. foldr (.) id (map (subGo x) z)
-}
where
go' = contentToBuilderRT cd render'
getMixin d =
case lookup d cd of
Nothing -> Left $ "Mixin not found: " ++ show d
Just (CDMixin m) -> Right m
Just _ -> Left $ "For " ++ show d ++ ", expected Mixin"
resolveAttr :: Attr Unresolved -> Either String (Attr Resolved)
resolveAttr (Attr k v) = Attr <$> (mconcat <$> mapM go' k) <*> (mconcat <$> mapM go' v)
subGo :: [Contents] -- ^ parent selectors
-> (HasLeadingSpace, Block Unresolved)
-> Either String (DList (Block Resolved))
subGo x' (hls, Block a b c d) =
blockRuntime cd render' (Block a' b c d)
where
a' = combineSelectors hls x' a
contentToBuilderRT :: [(Deref, CDData url)]
-> (url -> [(Text, Text)] -> Text)
-> Content
-> Either String Builder
contentToBuilderRT _ _ (ContentRaw s) = Right $ fromText $ pack s
contentToBuilderRT cd _ (ContentVar d) =
case lookup d cd of
Just (CDPlain s) -> Right s
_ -> Left $ show d ++ ": expected CDPlain"
contentToBuilderRT cd render' (ContentUrl d) =
case lookup d cd of
Just (CDUrl u) -> Right $ fromText $ render' u []
_ -> Left $ show d ++ ": expected CDUrl"
contentToBuilderRT cd render' (ContentUrlParam d) =
case lookup d cd of
Just (CDUrlParam (u, p)) ->
Right $ fromText $ render' u p
_ -> Left $ show d ++ ": expected CDUrlParam"
contentToBuilderRT _ _ ContentMixin{} = Left "contentToBuilderRT ContentMixin"
cssRuntime :: Bool -- ^ i2b?
-> Parser [TopLevel Unresolved]
-> FilePath
-> [(Deref, CDData url)]
-> (url -> [(Text, Text)] -> Text)
-> Css
cssRuntime toi2b parseBlocks fp cd render' = unsafePerformIO $ do
s' <- fmap TL.unpack $ qRunIO $ readUtf8File fp
let s = if toi2b then i2b s' else s'
let a = either (error . show) id $ parse parseBlocks s s
return $ CssWhitespace $ goTop [] a
where
goTop :: [(String, String)] -- ^ scope
-> [TopLevel Unresolved]
-> [TopLevel Resolved]
goTop _ [] = []
goTop scope (TopAtDecl dec cs':rest) =
TopAtDecl dec cs : goTop scope rest
where
cs = either error mconcat $ mapM (contentToBuilderRT cd render') cs'
goTop scope (TopBlock b:rest) =
map TopBlock (either error ($[]) $ blockRuntime (addScope scope) render' b) ++
goTop scope rest
goTop scope (TopAtBlock name s' b:rest) =
TopAtBlock name s (foldr (either error id . blockRuntime (addScope scope) render') [] b) :
goTop scope rest
where
s = either error mconcat $ mapM (contentToBuilderRT cd render') s'
goTop scope (TopVar k v:rest) = goTop ((k, v):scope) rest
addScope scope = map (DerefIdent . Ident *** CDPlain . fromString) scope ++ cd
vtToExp :: (Deref, VarType) -> Q Exp
vtToExp (d, vt) = do
d' <- lift d
c' <- c vt
return $ TupE [d', c' `AppE` derefToExp [] d]
where
c :: VarType -> Q Exp
c VTPlain = [|CDPlain . toCss|]
c VTUrl = [|CDUrl|]
c VTUrlParam = [|CDUrlParam|]
c VTMixin = [|CDMixin|]
getVars :: Monad m => [(String, String)] -> Content -> m [(Deref, VarType)]
getVars _ ContentRaw{} = return []
getVars scope (ContentVar d) =
case lookupD d scope of
Just _ -> return []
Nothing -> return [(d, VTPlain)]
getVars scope (ContentUrl d) =
case lookupD d scope of
Nothing -> return [(d, VTUrl)]
Just s -> fail $ "Expected URL for " ++ s
getVars scope (ContentUrlParam d) =
case lookupD d scope of
Nothing -> return [(d, VTUrlParam)]
Just s -> fail $ "Expected URLParam for " ++ s
getVars scope (ContentMixin d) =
case lookupD d scope of
Nothing -> return [(d, VTMixin)]
Just s -> fail $ "Expected Mixin for " ++ s
lookupD :: Deref -> [(String, b)] -> Maybe String
lookupD (DerefIdent (Ident s)) scope =
case lookup s scope of
Nothing -> Nothing
Just _ -> Just s
lookupD _ _ = Nothing
compressTopLevel :: TopLevel Unresolved
-> TopLevel Unresolved
compressTopLevel (TopBlock b) = TopBlock $ compressBlock b
compressTopLevel (TopAtBlock name s b) = TopAtBlock name s $ map compressBlock b
compressTopLevel x@TopAtDecl{} = x
compressTopLevel x@TopVar{} = x
compressBlock :: Block Unresolved
-> Block Unresolved
compressBlock (Block x y blocks mixins) =
Block (map cc x) (map go y) (map (second compressBlock) blocks) mixins
where
go (Attr k v) = Attr (cc k) (cc v)
cc [] = []
cc (ContentRaw a:ContentRaw b:c) = cc $ ContentRaw (a ++ b) : c
cc (a:b) = a : cc b
blockToMixin :: Name
-> Scope
-> Block Unresolved
-> Q Exp
blockToMixin r scope (Block _sel props subblocks mixins) =
[|Mixin
{ mixinAttrs = concat
$ $(listE $ map go props)
: map mixinAttrs $mixinsE
-- FIXME too many complications to implement sublocks for now...
, mixinBlocks = [] -- foldr (.) id $(listE $ map subGo subblocks) []
}|]
{-
. foldr (.) id $(listE $ map subGo subblocks)
. (concatMap mixinBlocks $mixinsE ++)
|]
-}
where
mixinsE = return $ ListE $ map (derefToExp []) mixins
go (Attr x y) = conE 'Attr
`appE` (contentsToBuilder r scope x)
`appE` (contentsToBuilder r scope y)
subGo (Block sel' b c d) = blockToCss r scope $ Block sel' b c d
blockToCss :: Name
-> Scope
-> Block Unresolved
-> Q Exp
blockToCss r scope (Block sel props subblocks mixins) =
[|((Block
{ blockSelector = $(selectorToBuilder r scope sel)
, blockAttrs = concat
$ $(listE $ map go props)
: map mixinAttrs $mixinsE
, blockBlocks = ()
, blockMixins = ()
} :: Block Resolved):)
. foldr (.) id $(listE $ map subGo subblocks)
. (concatMap mixinBlocks $mixinsE ++)
|]
where
mixinsE = return $ ListE $ map (derefToExp []) mixins
go (Attr x y) = conE 'Attr
`appE` (contentsToBuilder r scope x)
`appE` (contentsToBuilder r scope y)
subGo (hls, Block sel' b c d) =
blockToCss r scope $ Block sel'' b c d
where
sel'' = combineSelectors hls sel sel'
selectorToBuilder :: Name -> Scope -> [Contents] -> Q Exp
selectorToBuilder r scope sels =
contentsToBuilder r scope $ intercalate [ContentRaw ","] sels
contentsToBuilder :: Name -> Scope -> [Content] -> Q Exp
contentsToBuilder r scope contents =
appE [|mconcat|] $ listE $ map (contentToBuilder r scope) contents
contentToBuilder :: Name -> Scope -> Content -> Q Exp
contentToBuilder _ _ (ContentRaw x) =
[|fromText . pack|] `appE` litE (StringL x)
contentToBuilder _ scope (ContentVar d) =
case d of
DerefIdent (Ident s)
| Just val <- lookup s scope -> [|fromText . pack|] `appE` litE (StringL val)
_ -> [|toCss|] `appE` return (derefToExp [] d)
contentToBuilder r _ (ContentUrl u) =
[|fromText|] `appE`
(varE r `appE` return (derefToExp [] u) `appE` listE [])
contentToBuilder r _ (ContentUrlParam u) =
[|fromText|] `appE`
([|uncurry|] `appE` varE r `appE` return (derefToExp [] u))
contentToBuilder _ _ ContentMixin{} = error "contentToBuilder on ContentMixin"
type Scope = [(String, String)]
topLevelsToCassius :: [TopLevel Unresolved]
-> Q Exp
topLevelsToCassius a = do
r <- newName "_render"
lamE [varP r] $ appE [|CssNoWhitespace . foldr ($) []|] $ fmap ListE $ go r [] a
where
go _ _ [] = return []
go r scope (TopBlock b:rest) = do
e <- [|(++) $ map TopBlock ($(blockToCss r scope b) [])|]
es <- go r scope rest
return $ e : es
go r scope (TopAtBlock name s b:rest) = do
let s' = contentsToBuilder r scope s
e <- [|(:) $ TopAtBlock $(lift name) $(s') $(blocksToCassius r scope b)|]
es <- go r scope rest
return $ e : es
go r scope (TopAtDecl dec cs:rest) = do
e <- [|(:) $ TopAtDecl $(lift dec) $(contentsToBuilder r scope cs)|]
es <- go r scope rest
return $ e : es
go r scope (TopVar k v:rest) = go r ((k, v) : scope) rest
blocksToCassius :: Name
-> Scope
-> [Block Unresolved]
-> Q Exp
blocksToCassius r scope a = do
appE [|foldr ($) []|] $ listE $ map (blockToCss r scope) a
renderCss :: Css -> TL.Text
renderCss css =
toLazyText $ mconcat $ map go tops
where
(haveWhiteSpace, tops) =
case css of
CssWhitespace x -> (True, x)
CssNoWhitespace x -> (False, x)
go (TopBlock x) = renderBlock haveWhiteSpace mempty x
go (TopAtBlock name s x) =
fromText (pack $ concat ["@", name, " "]) `mappend`
s `mappend`
startBlock `mappend`
foldr mappend endBlock (map (renderBlock haveWhiteSpace (fromString " ")) x)
go (TopAtDecl dec cs) = fromText (pack $ concat ["@", dec, " "]) `mappend`
cs `mappend`
endDecl
startBlock
| haveWhiteSpace = fromString " {\n"
| otherwise = singleton '{'
endBlock
| haveWhiteSpace = fromString "}\n"
| otherwise = singleton '}'
endDecl
| haveWhiteSpace = fromString ";\n"
| otherwise = singleton ';'
renderBlock :: Bool -- ^ have whitespace?
-> Builder -- ^ indentation
-> Block Resolved
-> Builder
renderBlock haveWhiteSpace indent (Block sel attrs () ())
| null attrs = mempty
| otherwise = startSelect
<> sel
<> startBlock
<> mconcat (intersperse endDecl $ map renderAttr attrs)
<> endBlock
where
renderAttr (Attr k v) = startDecl <> k <> colon <> v
colon
| haveWhiteSpace = fromString ": "
| otherwise = singleton ':'
startSelect
| haveWhiteSpace = indent
| otherwise = mempty
startBlock
| haveWhiteSpace = fromString " {\n"
| otherwise = singleton '{'
endBlock
| haveWhiteSpace = fromString ";\n" `mappend` indent `mappend` fromString "}\n"
| otherwise = singleton '}'
startDecl
| haveWhiteSpace = indent `mappend` fromString " "
| otherwise = mempty
endDecl
| haveWhiteSpace = fromString ";\n"
| otherwise = singleton ';'
instance Lift Mixin where
lift (Mixin a b) = [|Mixin a b|]
instance Lift (Attr Unresolved) where
lift (Attr k v) = [|Attr k v :: Attr Unresolved |]
instance Lift (Attr Resolved) where
lift (Attr k v) = [|Attr $(liftBuilder k) $(liftBuilder v) :: Attr Resolved |]
liftBuilder :: Builder -> Q Exp
liftBuilder b = [|fromText $ pack $(lift $ TL.unpack $ toLazyText b)|]
instance Lift Content where
lift (ContentRaw s) = [|ContentRaw s|]
lift (ContentVar d) = [|ContentVar d|]
lift (ContentUrl d) = [|ContentUrl d|]
lift (ContentUrlParam d) = [|ContentUrlParam d|]
lift (ContentMixin m) = [|ContentMixin m|]
instance Lift (Block Unresolved) where
lift (Block a b c d) = [|Block a b c d|]
instance Lift (Block Resolved) where
lift (Block a b () ()) = [|Block $(liftBuilder a) b () ()|]
| psibi/shakespeare | Text/Css.hs | mit | 17,915 | 0 | 15 | 5,210 | 5,990 | 3,122 | 2,868 | -1 | -1 |
-- | A script for generating unrolled versions of three R2 benchmarks
-- For example, `makeCoinBias 42` will generate a file `CoinBias42.hs`
-- containing an unrolled version of the "coinBias" model that uses an
-- array of length 42
module Unroll where
import Text.PrettyPrint (parens, space, (<+>), punctuate, int, vcat,
Doc(..), text, render, ($$), (<>), nest)
-- Helpers
----------------------------------------------------------------------
makeNVars :: Int -> String -> [Doc]
makeNVars n var = [text var <> int i | i <- [0..n-1]]
makePair :: String -> Doc -> Doc -> Doc
makePair str a b = text str <+> a $$ b
pair :: Doc -> Doc -> Doc
pair = makePair "pair"
hPair :: Doc -> Doc -> Doc
hPair = makePair "HPair"
nested :: (Doc -> Doc -> Doc) -> [Doc] -> Doc
nested f vars = let ds = punctuate space vars
in foldr1 (\a -> parens . f a) ds
lastLine :: [Doc] -> String -> Doc
lastLine vars b = text "dirac" <+>
parens (pair (nested pair vars) (text b))
model :: Doc -> Doc -> Doc
model a b = text "Model" <+> a $$ b
makeNTypes :: Int -> String -> [Doc]
makeNTypes n typ = replicate n (text typ)
decl :: String -> Int -> String -> String -> Doc
decl name n a b =
let typeDecl = text name <+> text "::" <+>
model (nested hPair (makeNTypes n a))
(text b)
nameDecl = text name <+> text "="
in typeDecl $$ nameDecl
arrow :: Doc
arrow = text "->"
nat :: Int -> Doc
nat i = parens $ text "nat_" <+> int i
whereDef :: [String] -> Doc
whereDef defns = text "where" $$ vcat (map (nest l) (map text defns))
where l = length "where" + 1
-- Models
----------------------------------------------------------------------
coinBias :: Int -> Doc
coinBias n =
let firstTwo = decl "coinBias" n "HBool" "'HProb"
vars = makeNVars n "tossResult"
prog = vcat $
[text $ "beta (prob_ 2) (prob_ 5) >>= \\bias ->"] ++
[text "bern bias >>= \\" <> v <+> arrow | v <- vars] ++
[lastLine vars "bias"]
in firstTwo $$ nest 4 prog
digitRecognition :: Int -> Doc
digitRecognition n =
let firstTwo = decl "digitRecognition" n "HBool" "'HNat"
vars = makeNVars n "x"
prog = vcat $
[text $ "categorical dataPrior >>= \\y ->"] ++
[text "bern ((dataParams ! y) !" <+> nat i <> text ") >>= \\" <>
(vars !! i) <+> arrow | i <- [0..n-1]] ++
[lastLine vars "y"] ++
[whereDef ["dataPrior = var (Variable \"dataPrior\" 73 (SArray SProb))",
"dataParams = var (Variable \"dataParams\" 41 (SArray (SArray SProb)))"]]
in firstTwo $$ nest 4 prog
linearRegression :: Int -> Doc
linearRegression n =
let firstTwo = decl "linearRegression" n "'HReal" "HUnit"
vars = makeNVars n "y"
prog = vcat $
[text "normal (real_ 0) (prob_ 1) >>= \\a ->",
text "normal (real_ 5) (prob_ 1.825) >>= \\b ->",
text "gamma (prob_ 1) (prob_ 1) >>= \\invNoise ->"] ++
[text "normal (a * (dataX !" <+> nat i <>
text ")) (recip (sqrt invNoise)) >>= \\" <>
(vars !! i) <+> arrow | i <- [0..n-1]] ++
[lastLine vars "unit"] ++
[whereDef ["dataX = var (Variable \"dataX\" 73 (SArray SReal))"]]
in firstTwo $$ nest 4 prog
-- Make files
----------------------------------------------------------------------
pragmas :: Doc
pragmas = text "{-# LANGUAGE DataKinds, TypeOperators, OverloadedStrings #-}\n"
moduleName :: String -> Doc
moduleName name = text "module" <+> text name <+> text "where\n"
imports :: Doc
imports = vcat $
[text "import Prelude (print, length, IO)",
text "import Language.Hakaru.Syntax.Prelude",
text "import Language.Hakaru.Disintegrate",
text "import Language.Hakaru.Syntax.ABT",
text "import Language.Hakaru.Syntax.AST",
text "import Language.Hakaru.Types.DataKind",
text "import Language.Hakaru.Types.Sing\n"]
synonyms :: Doc
synonyms = text "type Model a b = TrivialABT Term '[] ('HMeasure (HPair a b))"
$$ text "type Cond a b = TrivialABT Term '[] (a ':-> 'HMeasure b)\n"
mainCall :: String -> Doc
mainCall name =
text "main :: IO ()" $$
text "main =" <+>
text "print (length (disintegrate" <+> text name <> text "))\n"
makeCoinBias :: Int -> IO ()
makeCoinBias n =
let name = "CoinBias" ++ show n
doc = pragmas $$
moduleName name $$
imports $$
synonyms $$
coinBias n <> text "\n" $$
mainCall "coinBias"
in writeFile (name ++ ".hs") (render doc)
makeDigitRecognition :: Int -> IO ()
makeDigitRecognition n =
let name = "DigitRecognition" ++ show n
doc = pragmas $$
moduleName name $$
imports $$
synonyms $$
digitRecognition n <> text "\n" $$
mainCall "digitRecognition"
in writeFile (name ++ ".hs") (render doc)
makeLinearRegression :: Int -> IO ()
makeLinearRegression n =
let name = "LinearRegression" ++ show n
doc = pragmas $$
moduleName name $$
imports $$
synonyms $$
linearRegression n <> text "\n" $$
mainCall "linearRegression"
in writeFile (name ++ ".hs") (render doc)
main :: IO ()
main = makeCoinBias 5 >>
makeCoinBias 500
| zaxtax/hakaru | haskell/Tests/Unroll/Unroll.hs | bsd-3-clause | 5,683 | 0 | 17 | 1,807 | 1,546 | 782 | 764 | 124 | 1 |
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE GADTs #-}
{-# LANGUAGE PolyKinds #-}
{-# LANGUAGE TemplateHaskell #-}
{-# LANGUAGE TypeFamilies #-}
module T17296 where
import Data.Foldable
import Data.Kind
import Language.Haskell.TH hiding (Type)
import System.IO
data family Foo1 :: Type -> Type
data instance Foo1 Bool = Foo1Bool
data instance Foo1 (Maybe a)
data family Foo2 :: k -> Type
data instance Foo2 Bool = Foo2Bool
data instance Foo2 (Maybe a)
data instance Foo2 :: Char -> Type
data instance Foo2 :: (Char -> Char) -> Type where
data family Foo3 :: k
data instance Foo3
data instance Foo3 Bool = Foo3Bool
data instance Foo3 (Maybe a)
data instance Foo3 :: Char -> Type
data instance Foo3 :: (Char -> Char) -> Type where
$(do let test :: Name -> Q ()
test n = do i <- reify n
runIO $ do hPutStrLn stderr $ pprint i
hPutStrLn stderr ""
hFlush stderr
traverse_ test [''Foo1, ''Foo2, ''Foo3]
pure [])
| sdiehl/ghc | testsuite/tests/th/T17296.hs | bsd-3-clause | 1,017 | 6 | 10 | 274 | 297 | 171 | 126 | -1 | -1 |
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="hu-HU">
<title>Alert Filters | ZAP Extension</title>
<maps>
<homeID>top</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>Contents</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>Index</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Keresés</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Favorites</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | denniskniep/zap-extensions | addOns/alertFilters/src/main/javahelp/org/zaproxy/zap/extension/alertFilters/resources/help_hu_HU/helpset_hu_HU.hs | apache-2.0 | 976 | 78 | 66 | 159 | 416 | 210 | 206 | -1 | -1 |
module SortProps where
import Sort
assert S0 = {sort []} ::: []
assert S1 = All x . {sort [x]} === {[x]}
assert S2 = All x1 . All x2 . {sort [x1,x2]} === {[x1,x2]}
\/ {sort [x1,x2]} === {[x2,x1]}
--assert S = S0 /\ S1 /\ S2
ordered [] = True
ordered [x] = True
ordered (x1:x2:xs) = x1<=x2 && ordered (x2:xs)
lteAll x = all (x<=)
orderedInt :: [Int] -> Bool
orderedInt = ordered
--property IsTrue = {| x | x===True |}
--property IsOrdered = {|xs | IsTrue {orderedInt xs} |}
property IsOrdered = !ordered
assert InsertProp = All x . All xs . IsOrdered xs ==> IsOrdered {insert x xs}
assert SortProp1 = All xs . IsOrdered {sort xs}
property AllElems P = Gfp X . [] \/ P:X
property Minimal x = AllElems (!((x::Int)<=))
property Or1 P Q = {| x | (x:::P) \/ (x:::Q) |}
property Or2 P Q = P \/ Q
property IsOrdered2 =
Lfp X . {| xs | (xs:::[]) \/ (Minimal {head xs} xs /\ ({tail xs}:::X)) |}
| forste/haReFork | tools/hs2alfa/tests/SortProps.hs | bsd-3-clause | 931 | 25 | 8 | 223 | 418 | 246 | 172 | -1 | -1 |
-- Test the classic "\SOH" ambiguity
module Main(main) where
main = do { print soh ; print (length (fst (head soh))) ;
print so ; print (length (fst (head so))) }
where
so, soh :: [(String,String)]
soh = reads "\"\\SOH\"" -- Should read \SOH
so = reads "\"\\SOx\"" -- Should read \SO followed by x
| ezyang/ghc | libraries/base/tests/reads001.hs | bsd-3-clause | 343 | 0 | 13 | 100 | 116 | 63 | 53 | 6 | 1 |
-- Check that the record selector for maskMB unfolds in the body of f
-- At one stage it didn't because the implicit unfolding looked too big
-- Trac #2581
module ShouldCompile where
import Data.Array.Base
data MBloom s a = MB {
shiftMB :: {-# UNPACK #-} !Int
, maskMB :: {-# UNPACK #-} !Int
, bitArrayMB :: {-# UNPACK #-} !(STUArray s Int Int)
}
f a b c = case maskMB (MB a b c) of
3 -> True
_ -> False
| bitemyapp/ghc | testsuite/tests/eyeball/record1.hs | bsd-3-clause | 453 | 0 | 11 | 133 | 99 | 57 | 42 | 9 | 2 |
{-# OPTIONS_GHC -F -pgmF hspec-discover -optF --no-main #-}
| AndrewRademacher/twitter-conduit | tests/Spec.hs | bsd-2-clause | 60 | 0 | 2 | 8 | 3 | 2 | 1 | 1 | 0 |
-- Test for trac #2141
module Foo where
foo :: () -> ()
foo x = x { foo = 1 }
| wxwxwwxxx/ghc | testsuite/tests/rename/should_fail/rnfail054.hs | bsd-3-clause | 80 | 0 | 6 | 24 | 34 | 20 | 14 | 3 | 1 |
module Annfail13 where
-- Testing that brackets are mandatory in the ANN syntax
{-# ANN f id 1 #-}
{-# ANN f 1 :: Int #-}
f x = x | urbanslug/ghc | testsuite/tests/annotations/should_fail/annfail13.hs | bsd-3-clause | 130 | 0 | 5 | 30 | 15 | 10 | 5 | -1 | -1 |
module Main where
import Flight
main = do
input <- readFile "input.txt"
let constraints = map parseConstraint (lines input)
print $ maxScore 2503 constraints
| corajr/adventofcode2015 | 14/Main.hs | mit | 166 | 0 | 12 | 32 | 54 | 26 | 28 | 6 | 1 |
module Rebase.GHC.MVar
(
module GHC.MVar
)
where
import GHC.MVar
| nikita-volkov/rebase | library/Rebase/GHC/MVar.hs | mit | 68 | 0 | 5 | 12 | 20 | 13 | 7 | 4 | 0 |
insert x [] = [x]
insert x (y:ys)
| x < y = x:y:ys
| otherwise = y : insert x ys
isort [] = []
isort (x:xs) = insert x (isort xs)
main = do
print $ isort [4, 6, 9, 8, 3, 5, 1, 7, 2]
| shigemk2/haskell_abc | insert.hs | mit | 203 | 0 | 9 | 69 | 149 | 77 | 72 | 8 | 1 |
module Options
( Command(..)
, Options(..)
, BuildParams(..)
, AuthCreds(..)
, parseOptions
) where
import Options.Applicative
import qualified Data.Text as T
import qualified Data.ByteString.Char8 as BS
import Data.Maybe (mapMaybe)
import Jenkins.Types
type JobId = T.Text
newtype BuildParams = BuildParams {
fromBuildParams :: [(BS.ByteString, BS.ByteString)]
} deriving (Show, Eq)
newtype AuthCreds = AuthCreds {
fromAuthCreds :: (BS.ByteString, BS.ByteString)
} deriving (Show, Eq)
data Command = JobStatuses
| JobStatus JobId
| RunBuild JobId BuildParams
| BuildLog JobId (Maybe BuildNum)
deriving (Show, Eq)
data Options = Options
{ optsBaseUri :: String
, optsAuth :: Maybe AuthCreds
, optsCommand :: Command
} deriving (Show, Eq)
parserInfo :: Parser Command -> String -> ParserInfo Command
parserInfo cmd desc = info (helper <*> cmd) (progDesc desc)
jobIdParser :: String -> ReadM JobId
jobIdParser = return . T.pack
buildParamParser :: String -> ReadM BuildParams
buildParamParser = return . BuildParams . mapMaybe parseParam . BS.words . BS.pack
parseParam :: BS.ByteString -> Maybe (BS.ByteString, BS.ByteString)
parseParam s =
case (BS.break ((==) '=') s) of
(_, "") -> Nothing
(k, v) -> Just (k, BS.drop 1 v)
authCredsParser :: String -> ReadM (Maybe AuthCreds)
authCredsParser s = do
return $ case BS.splitWith ((==) ':') (BS.pack s) of
(user:pass:[]) -> Just (AuthCreds (user, pass))
_ -> Nothing
buildNumParser :: String -> ReadM (Maybe BuildNum)
buildNumParser = pure . Just . BuildNum . read
jobStatusParser :: Parser Command
jobStatusParser = JobStatus
<$> argument (str >>= jobIdParser) ( metavar "JOB_ID" )
runBuildParser :: Parser Command
runBuildParser = RunBuild
<$> argument (str >>= jobIdParser) ( metavar "JOB_ID" )
<*> argument (str >>= buildParamParser) ( metavar "PARAM=VALUE .."
<> value (BuildParams [])
<> help "List of parameter/value pairs"
)
buildLogParser :: Parser Command
buildLogParser = BuildLog
<$> argument (str >>= jobIdParser) ( metavar "JOB_ID" )
<*> argument (str >>= buildNumParser) ( metavar "BUILD_NUM"
<> value Nothing
<> help "Build number"
)
parseOptions :: Parser Options
parseOptions = Options
<$> strOption ( short 's'
<> metavar "JENKINS_URL"
<> help "Jenkins base URL"
)
<*> option (str >>= authCredsParser ) ( short 'u'
<> metavar "HTTP_AUTH"
<> value Nothing
<> help "http authentication credentials (i.e. user:password)"
)
<*> subparser
( command "jobs" jobStatusesParserInfo
<> command "job" jobStatusParserInfo
<> command "build" runBuildParserInfo
<> command "log" buildLogParserInfo
)
jobStatusesParserInfo :: ParserInfo Command
jobStatusesParserInfo =
parserInfo (pure JobStatuses) "display all jobs' status"
jobStatusParserInfo :: ParserInfo Command
jobStatusParserInfo =
parserInfo jobStatusParser "list recent builds for a given job"
runBuildParserInfo :: ParserInfo Command
runBuildParserInfo =
parserInfo runBuildParser "build a given job"
buildLogParserInfo :: ParserInfo Command
buildLogParserInfo =
parserInfo buildLogParser "stream build log to standard output"
| afiore/jenkins-tty.hs | src/Options.hs | mit | 3,617 | 0 | 14 | 1,002 | 964 | 512 | 452 | 87 | 2 |
import Control.Monad.Reader.Class (ask)
import XMonad.Main (xmonad)
import XMonad.Core
( X
, spawn
, ManageHook
, layoutHook
, startupHook
, manageHook
, logHook
, modMask
, terminal
, borderWidth
, normalBorderColor
, focusedBorderColor
, workspaces
)
import XMonad.ManageHook
( composeAll
, (<||>), (=?), (-->), (<+>)
, className, doF, stringProperty )
import XMonad.Config ( defaultConfig )
import XMonad.Hooks.DynamicLog
( dynamicLogWithPP
, ppOutput
, ppTitle
, ppLayout
, ppVisible
, ppCurrent
, xmobarPP
, xmobarColor
, shorten
)
import XMonad.Hooks.ManageDocks ( avoidStruts, docks, manageDocks)
import XMonad.Util.Run (spawnPipe)
import XMonad.Util.EZConfig (additionalKeys)
import XMonad.Layout.Spacing (smartSpacing)
import XMonad.Layout.NoBorders (smartBorders)
import XMonad.StackSet (sink)
import Graphics.X11.Types
import Data.Bits ( (.|.) )
import System.IO (hPutStrLn)
----------------------- Colors ----------------------
magenta = "#FF14E5"
blue1 = "#29acff"
green1 = "#60ff45"
white1 = "#FFFFFF"
gray1 = "#3D3D3D"
gray2 = "#808080"
gray3 = "#CCCCCC"
yellow1 = "#FFF500"
yellow2 = "#ffff66"
myManageHook :: ManageHook
myManageHook = composeAll [ (role =? "gimp-toolbox" <||> role =? "gimp-image-window") -->(unfloat)
, className =? "MPlayer" --> (unfloat) ]
where unfloat = ask >>= doF . sink
role = stringProperty "WM_WINDOW_ROLE"
myLayoutHook = smartSpacing 5 $ avoidStruts $ layoutHook defaultConfig
myKeys = [ ((mod4Mask .|. shiftMask, xK_z), spawn "xscreensaver-command -lock")
, ((mod4Mask .|. shiftMask, xK_o), spawn "urxvt -e ranger")
, ((mod4Mask .|. shiftMask, xK_f), spawn "firefox")
, ((controlMask, xK_Print), spawn "sleep 0.2; scrot -s")
, ((0, xK_Print), spawn "scrot")
]
myLogHook h = dynamicLogWithPP xmobarPP { ppOutput = hPutStrLn h
, ppTitle = xmobarColor blue1 "" . shorten 70
, ppCurrent = xmobarColor blue1 ""
, ppVisible = xmobarColor white1 ""
--, ppHiddenNoWindows = xmobarColor magenta ""
, ppLayout = xmobarColor gray2 ""
}
------------------------ Bottom bar stuff ----------------------
myStartupHook :: X ()
myStartupHook = do spawn xmobarBottom
bin_xmobar = "/run/current-system/sw/bin/xmobar"
rc_xmobarTop = "/home/bernie/.xmobarrc.hs"
rc_xmobarBottom = "/home/bernie/.xmobarrc_bottom.hs"
xmobarTop = bin_xmobar ++ " " ++ rc_xmobarTop
xmobarBottom = bin_xmobar ++ " " ++ rc_xmobarBottom
------------------------------------------------------------------
main :: IO ()
main = do
xmproc <- spawnPipe xmobarTop
xmonad $ docks defaultConfig
{ manageHook = manageDocks <+> myManageHook <+> manageHook defaultConfig
, layoutHook = myLayoutHook
, logHook = myLogHook xmproc
-- , startupHook = myStartupHook
, modMask = mod4Mask
, terminal = "urxvt"
, borderWidth = 1
, normalBorderColor = gray1
, focusedBorderColor = blue1
, workspaces = ["sh1","sh2","sh3","gimp4","pdf5","com6","mpc7","web8","web9"]
} `additionalKeys` myKeys
| iambernie/dotfiles | .xmonad/xmonad.hs | mit | 3,536 | 0 | 13 | 1,031 | 776 | 462 | 314 | 86 | 1 |
sumSquareDifference :: Int
sumSquareDifference =
let boundary = 100
sumSquare = sum [i^2 | i <- [1..100]]
squareSum = sum [1..100] ^ 2
in squareSum - sumSquare
| samidarko/euler | problem006.hs | mit | 187 | 0 | 13 | 54 | 71 | 37 | 34 | 6 | 1 |
import qualified Data.Map as Map
data Person = Person { firstName :: String
, lastName :: String
, age :: Int
, height :: Float
, phoneNumber :: String
, flavor :: String } deriving (Show)
data Car = Car { company :: String
, model :: String
, year :: Int } deriving (Show)
{- let car = Car { company="TOYOTA", model="Lexus", year=2015 } -}
tellCar :: Car -> String
tellCar (Car { company = c, model = m, year = y })
= "This " ++ c ++ " " ++ m ++ " was made in " ++ show y
data Vector a = Vector a a a deriving (Show)
vplus :: (Num a) => Vector a -> Vector a -> Vector a
(Vector i j k) `vplus` (Vector l m n)
= Vector (i+l) (j+m) (k+n)
dotProd :: (Num a) => Vector a -> Vector a -> a
(Vector i j k) `dotProd` (Vector l m n)
= i*l + j*m + k*n
vmult :: (Num a) => Vector a -> a -> Vector a
(Vector i j k) `vmult` m = Vector (i*m) (j*m) (k*m)
data UniquePerson = UniquePerson { u_firstName :: String
, u_lastName :: String
, u_age :: Int } deriving (Eq, Show, Read)
data Day = Sunday | Monday | Tuesday | Wednesday | Thursday | Friday | Saturday
deriving (Eq, Ord, Show, Read, Bounded, Enum)
type PhoneNumber = String
type Name = String
type PhoneBook = [(Name, PhoneNumber)]
phoneBook :: PhoneBook
phoneBook
= [("betty", "111-2222")
,("bonnie", "222-3333")
,("patsy", "333-4444")
,("lucille", "444-5555")
,("wendy", "555-6666")
,("penny", "666-7777")]
inPhoneBook :: Name -> PhoneNumber -> PhoneBook -> Bool
inPhoneBook name pnumber pbook = (name, pnumber) `elem` pbook
data LockerState = Taken | Free
deriving (Show, Eq)
type Code = String
type LockerMap = Map.Map Int (LockerState, Code)
lockerLookup :: Int -> LockerMap -> Either String Code
lockerLookup lockerNumber map = case Map.lookup lockerNumber map of
Nothing -> Left $ "Locker " ++ show lockerNumber ++ " doesn't exist!"
Just (state, code) -> if state /= Taken
then Right code
else Left $ "Locker " ++ show lockerNumber ++ " is already taken!"
lockers :: LockerMap
lockers = Map.fromList
[(100, (Taken, "AB111"))
,(101, (Free, "BC222"))
,(102, (Free, "CD333"))
,(105, (Free, "DE444"))
,(107, (Taken, "EF555"))
,(109, (Taken, "FG666"))]
{- data List a = Empty | Cons a (List a) -}
{- deriving (Show, Read, Eq, Ord) -}
infixr 5 ^++
(^++) :: [a] -> [a] -> [a]
[] ^++ ys = ys
(x:xs) ^++ ys = x:(xs ^++ ys)
data Tree a = EmptyTree | Node a (Tree a) (Tree a) deriving (Show)
singleton :: a -> Tree a
singleton x = Node x EmptyTree EmptyTree
treeInsert :: (Ord a) => a -> Tree a -> Tree a
treeInsert x EmptyTree = singleton x
treeInsert x (Node a left right)
| x == a = Node x left right
| x < a = Node a (treeInsert x left) right
| x > a = Node a left (treeInsert x right)
treeElem :: (Ord a) => a -> Tree a -> Bool
treeElem x EmptyTree = False
treeElem x (Node a left right)
| x == a = True
| x < a = treeElem x left
| x > a = treeElem x right
{- let nums = [8,6,4,1,7,3,5] -}
{- let numsTree = foldr treeInsert EmptyTree nums -}
{- numsTree -}
{- 8 `treeElem` numsTree -}
{- class Eq a where -}
{- (==) :: a -> a -> Bool -}
{- (/=) :: a -> a -> Bool -}
{- x == y = not (x /= y) -}
{- x /= y = not (x == y) -}
data TrafficLight = Red | Yellow | Green
instance Eq TrafficLight where
Red == Red = True
Green == Green = True
Yellow == Yellow = True
_ == _ = False
instance Show TrafficLight where
show Red = "Red Light"
show Yellow = "Yellow Light"
show Green = "Green Light"
class YesNo a where
yesno :: a -> Bool
instance YesNo Int where
yesno 0 = False
yesno _ = True
instance YesNo [a] where
yesno [] = False
yesno _ = True
instance YesNo Bool where
yesno = id
instance YesNo [Maybe a] where
yesno (Just _) = True
yesno Nothing = False
instance YesNo (Tree a) where
yesno EmptyTree = False
yesno _ = True
instance YesNo TrafficLight where
yesno Red = False
yesno _ = True
{- yesno $ length [] -}
{- yesno $ "haha" -}
{- yesno $ "" -}
{- yesno $ Just 0 -}
{- yesno $ Tree -}
{- yesno EmptyTree -}
{- yesno [] -}
{- yesno [0,0,0] -}
yesnoIf :: (YesNo y) => y -> a -> a -> a
yesnoIf yesnoVal yesResult noResult
= if yesno yesnoVal
then yesResult
else noResult
{- yesnoIf [] "YEAH!" "NO!" -}
{- class Functor f where -}
{- fmap :: (a -> b) -> f a -> f b -}
{- instance Functor [] where -}
{- fmap = map -}
{- instance Functor Maybe where -}
{- fmap f (Just x) = Just (f x) -}
{- fmap f Nothing = Nothing -}
| yhoshino11/learning_haskell | ch7/ch7.hs | mit | 4,769 | 3 | 11 | 1,377 | 1,672 | 905 | 767 | 111 | 3 |
{-# LANGUAGE MultiWayIf #-}
{-# LANGUAGE TypeSynonymInstances #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE FlexibleContexts #-}
module Language.Jass.Codegen.Type(
toLLVMType
, toLLVMType'
, defaultValue
, jassArray
, arraySize
, sizeOfType
, codeTypeSize
, codeTypeStruct
, pointerSize
, getFunctionType
, getFunctionArgumentsTypes
, getFunctionReturnType
, isIntegralType
, isStringType
, getReference
, getCallableLLVMType
, getTypeId
, getTypeFromId
, TypesMap
, getTypeMap
, module SemType
) where
import Language.Jass.JassType
import Language.Jass.Parser.AST.Parameter as AST
import Language.Jass.Codegen.Context
import Language.Jass.Semantic.Type as SemType
import LLVM.General.AST as LLVM
import LLVM.General.AST.Type as LLVMType
import LLVM.General.AST.Constant as LLVM
import LLVM.General.AST.Float
import LLVM.General.AST.DataLayout
import LLVM.General.AST.AddrSpace
import Control.Arrow
import Control.Monad.Except
import Language.Jass.Semantic.Callable
import Language.Jass.Semantic.Variable
import qualified Data.Map.Lazy as ML
import qualified Data.HashMap.Strict as HM
-- | Default array size
arraySize :: Num a => a
arraySize = 65536
-- | Returns reference to local or global variable
getReference :: String -> Codegen (LLVM.Type, LLVM.Operand)
getReference name = do
mvar <- getVariable name
case mvar of
Just var -> do
varType <- ptr <$> if isVarArray var then toLLVMType (JArray $ getVarType var) else toLLVMType (getVarType var)
return (varType, if isGlobalVariable var then LLVM.ConstantOperand $ LLVM.GlobalReference varType (LLVM.Name name) else LLVM.LocalReference varType (LLVM.Name name))
Nothing -> throwError $ unplacedSemError $ "ICE: cannot find variable " ++ name
-- | Returns callable llvm type
getCallableLLVMType :: Callable -> Codegen LLVM.Type
getCallableLLVMType callable = do
rt <- toLLVMType' $ getCallableReturnType callable
args <- mapM toLLVMType (getParamType <$> getCallableParameters callable)
return FunctionType {
resultType = rt
, argumentTypes = args
, isVarArg = False
}
-- | Converts jass type to LLVM type
toLLVMType :: JassType -> Codegen Type
toLLVMType JInteger = return i32
toLLVMType JReal = return float
toLLVMType JBoolean = return i1
toLLVMType JString = return $ ptr i8
toLLVMType JHandle = return i64
toLLVMType JCode = return $ ptr codeTypeStruct
toLLVMType (JArray et) = ArrayType arraySize <$> toLLVMType et
toLLVMType t@(JUserDefined _) = toLLVMType =<< getRootType t
toLLVMType JNull = throwError $ unplacedSemError "ICE: cannot generate code for special type JNull"
sizeOfType :: JassType -> Codegen Int
sizeOfType JInteger = return 4
sizeOfType JReal = return 4
sizeOfType JBoolean = return 1
sizeOfType JString = return pointerSize
sizeOfType JHandle = return 8
sizeOfType JCode = return codeTypeSize
sizeOfType (JArray et) = (arraySize *) <$> sizeOfType et
sizeOfType t@(JUserDefined _) = sizeOfType =<< getRootType t
sizeOfType JNull = throwError $ unplacedSemError "ICE: cannot generate code for special type JNull"
-- | Internal representation of code value
codeTypeStruct :: Type
codeTypeStruct = StructureType {
LLVMType.isPacked = False,
elementTypes = [
ptr i8 -- function pointer
, i32 -- id of return type, 0 for nothing
, i32 -- count of arguments
, ptr i32 -- array of arguments types as ids
]
}
-- | Returns size of code value internal representation
codeTypeSize :: Int
codeTypeSize = pointerSize + 4 + 4 + pointerSize
-- | Size of pointer in jass code
pointerSize :: Int
pointerSize = fromIntegral $ fst (pointerLayouts jassDataLayout ML.! AddrSpace 0)
-- | Ditto, including void type
toLLVMType' :: Maybe JassType -> Codegen Type
toLLVMType' = maybe (return VoidType) toLLVMType
-- | Returns default value for a type
defaultValue :: JassType -> Codegen Constant
defaultValue JInteger = return $ Int 32 0
defaultValue JReal = return $ Float (Single 0.0)
defaultValue JBoolean = return $ Int 1 0
defaultValue JString = return $ Null (ptr i8)
defaultValue JHandle = return $ Int 64 0
defaultValue JCode = throwError $ unplacedSemError "ICE: no default value for code value"
defaultValue t@(JArray _) = Null <$> toLLVMType t
defaultValue t@(JUserDefined _) = defaultValue =<< getRootType t
defaultValue JNull = throwError $ unplacedSemError "ICE: cannot generate code for special type JNull"
-- | Generates array type from element LLVM type
jassArray :: Type -> Type
jassArray = ArrayType arraySize
-- | Returns LLVM type of a function
getFunctionType :: String -> Codegen LLVM.Type
getFunctionType name = do
callable <- getCallable name
case callable of
Nothing -> throwError $ unplacedSemError $ "ICE: cannot find function " ++ name
Just fn -> do
retType <- toLLVMType' $ getCallableReturnType fn
pars <- mapM convertPars $ getCallableParameters fn
return $ FunctionType retType pars False
where
convertPars (AST.Parameter _ t _) = toLLVMType t
-- | Converts callable arguments types to LLVM types
getFunctionArgumentsTypes :: String -> Codegen [LLVM.Type]
getFunctionArgumentsTypes name = do
callable <- getCallable name
case callable of
Nothing -> throwError $ unplacedSemError $ "ICE: cannot find function " ++ name
Just fn -> mapM convertPars $ getCallableParameters fn
where
convertPars (AST.Parameter _ t _) = toLLVMType t
-- | Returns function return type in LLVM typesystem
getFunctionReturnType :: String -> Codegen LLVM.Type
getFunctionReturnType name = do
callable <- getCallable name
case callable of
Nothing -> throwError $ unplacedSemError $ "ICE: cannot find function " ++ name
Just fn -> maybe (return VoidType) toLLVMType $ getCallableReturnType fn
-- | Returns true if type is some sort of integer
isIntegralType :: LLVM.Type -> Bool
isIntegralType (IntegerType _) = True
isIntegralType _ = False
-- | Returns true if type is represents string type
isStringType :: LLVM.Type -> Bool
isStringType (PointerType (IntegerType 8) _) = True
isStringType _ = False
-- | Returns jass type id, custom types should be registered before the function is called
getTypeId :: Maybe JassType -> Codegen Int
getTypeId Nothing = return 0
getTypeId (Just JInteger) = return 1
getTypeId (Just JReal) = return 2
getTypeId (Just JBoolean) = return 3
getTypeId (Just JString) = return 4
getTypeId (Just JHandle) = return 5
getTypeId (Just JCode) = return 6
getTypeId (Just (JArray et)) = (256 +) <$> getTypeId (Just et)
getTypeId (Just (JUserDefined n)) = (512 +) <$> getCustomTypeId n
getTypeId (Just JNull) = throwError $ unplacedSemError "ICE: cannot generate code for special type JNull"
-- | Returns jass type by runtime id, custom types should be registered before the function is called
getTypeFromId :: Int -> Codegen (Maybe JassType)
getTypeFromId 0 = return Nothing
getTypeFromId 1 = return $ Just JInteger
getTypeFromId 2 = return $ Just JReal
getTypeFromId 3 = return $ Just JBoolean
getTypeFromId 4 = return $ Just JString
getTypeFromId 5 = return $ Just JHandle
getTypeFromId 6 = return $ Just JCode
getTypeFromId n
| n > 512 = Just . JUserDefined <$> getCustomTypeFromId (n - 512)
| n > 256 = fmap JArray <$> getTypeFromId (n - 256)
| otherwise = throwError $ unplacedSemError $ "ICE: unknown id of type '" ++ show n ++ "'"
type TypesMap = (HM.HashMap Int JassType, HM.HashMap JassType Int)
getTypeMap :: Codegen TypesMap
getTypeMap = do
let ts = [JInteger, JReal, JBoolean, JString, JHandle, JCode]
basic <- mapM getTypeId (fmap Just ts)
custom <- second (HM.fromList . fmap (first JUserDefined) . HM.toList) <$> first (fmap JUserDefined) <$> getCustomTypes
return $ (
HM.fromList (basic `zip` ts) `HM.union` (fst custom),
HM.fromList (ts `zip` basic) `HM.union` (snd custom)) | NCrashed/hjass | src/library/Language/Jass/Codegen/Type.hs | mit | 7,842 | 0 | 19 | 1,392 | 2,153 | 1,108 | 1,045 | 167 | 4 |
{-# LANGUAGE RecordWildCards #-}
module Interactive.Pipeline
( run_pipeline, Params' (..), Params, pos )
where
-- Modules
--import Browser
import Document.Document
import Documentation.SummaryGen
import Utilities.Config hiding ( wait )
import Interactive.Observable
import Interactive.Serialize
import Logic.Expr
import UnitB.UnitB
import Z3.Z3
( discharge
, Validity ( .. ) )
-- Libraries
import Control.DeepSeq
import Control.Concurrent
import Control.Concurrent.STM
import Control.Lens
import Control.Exception
import Control.Monad
import Control.Monad.Trans
import Control.Monad.Trans.Either
import Control.Monad.Trans.State
import Control.Precondition
import Data.Char
import qualified Data.List as L
import Data.Map as M
( insert, keys
, toList, unions )
import qualified Data.Map as M
import GHC.Generics (Generic)
import System.Console.ANSI
import System.Directory
import System.Process
import Text.Printf.TH
import Utilities.Syntactic
import Utilities.TimeIt
-- The pipeline is made of three processes:
-- o the parser
-- o the prover
-- o the display
--
-- The prover and the parser _share_ a map of proof obligations
-- The prover and the parser _share_ a list of PO labels
-- The
data Shared = Shared
{ working :: Observable Int
, system :: Observable System
, error_list :: Observable [Error]
, pr_obl :: Observable (M.Map Key (Seq,Maybe Bool))
, fname :: FilePath
, exit_code :: MVar ()
, parser_state :: Observable ParserState
, focus :: Observable (Maybe String)
, dump_cmd :: Observable (Maybe DumpCmd)
, redraw :: Observable Bool
}
data ParserState = Idle Double | Parsing
deriving Eq
type Params = Params' (M.Map Label (M.Map Label (Bool,Seq)))
data Params' pos = Params
{ path :: FilePath
, verbose :: Bool
, continuous :: Bool
, no_dump :: Bool
, no_verif :: Bool
, reset :: Bool
, _pos :: pos
, init_focus :: Maybe String
} deriving (Generic)
makeLenses ''Params'
instance NFData Params where
instance Show ParserState where
show (Idle x) = [s|Idle %sms|] $ show $ round $ x * 1000
show Parsing = "Parsing"
parser :: Shared
-> IO (IO ())
parser (Shared { .. }) = return $ do
t <- getModificationTime fname
write_obs parser_state Parsing
(dt,()) <- timeItT $ parse
write_obs parser_state (Idle dt)
evalStateT (forever $ do
liftIO $ do
threadDelay 250000
t0 <- get
t1 <- liftIO $ getModificationTime fname
if t0 == t1 then return ()
else do
put t1
liftIO $ do
write_obs parser_state Parsing
(t,()) <- timeItT parse
write_obs parser_state (Idle t)
) t
where
f m = return $ M.mapKeys (g $ _name m) $ proof_obligation m
-- return $ fromList $ map (g $ _name m) $ toList $ x
g lbl x = (lbl,x)
h lbl (x,y) = ((lbl,x),y)
parse = do
xs <- liftIO $ runEitherT $ do
s <- EitherT $ parse_system fname
ms <- hoistEither $ mapM f $ M.elems $ s!.machines
pos <- hoistEither $ mapM theory_po $ M.elems $ s!.theories
let cs = M.fromList $ map (uncurry h) $ do
(x,ys) <- zip (map label (s!.theories.to keys)) pos
y <- toList ys
return (x,y)
liftIO $ evaluate (ms, cs, s)
case xs of
Right (ms,cs,s) -> do
let new_pos = unions (cs : map (M.mapKeys $ over _1 as_label) ms) :: M.Map Key Seq
f (s0,b0) (s1,b1)
| s0 == s1 = (s0,b0)
| otherwise = (s1,b1)
g s = (s, Nothing)
write_obs_fast system s
write_obs error_list []
modify_obs_fast pr_obl $ \pos -> do
evaluate $ M.unionWith f (pos `M.intersection` new_pos) (M.map g new_pos)
return ()
Left es -> do
write_obs error_list es
return ()
prover :: Shared -> IO (IO ())
prover (Shared { .. }) = do
tok <- newEmptyMVar
observe pr_obl tok
-- req <- newEmptyMVar
req <- newTBQueueIO 20
forM_ [1..40] $ \p -> forkOn p $ worker req
return $ forever $ do
takeMVar tok
inc 1
po <- read_obs pr_obl
forM_ (keys po) $ \k -> do
po <- reads_obs pr_obl $ M.lookup k
case po of
Just (po,Nothing) -> do
liftIO $ atomically $ writeTBQueue req (k,po)
-- liftIO $ putMVar req (k,po)
_ -> return ()
dec 1
where
inc x = modify_obs working (return . (+x))
dec x = modify_obs working (return . (+ (-x)))
-- handler ::
handler lbl@(_,x) (ErrorCall msg) = do
write_obs dump_cmd $ Just $ Only x
fail ([s|During %s: %s|] (pretty lbl) msg)
worker req = forever $ do
-- (k,po) <- takeMVar req
(k,po) <- atomically $ readTBQueue req
let k' = uncurry (</>) k
inc 1
r <- catch (discharge k' po) (handler k)
dec 1
modify_obs pr_obl $ return . insert k (po,Just $ r == Valid)
proof_report :: Maybe String
-> M.Map Key (Seq,Maybe Bool)
-> [Error] -> Bool
-> [String]
proof_report = proof_report' False
proof_report' :: Bool
-> Maybe String
-> M.Map Key (Seq,Maybe Bool)
-> [Error] -> Bool
-> [String]
proof_report' showSuccess pattern outs es isWorking =
header ++
ys ++
( if null es then []
else "> errors" : map report es ) ++
footer ++
[ if isWorking
then "> working ..."
else " "
]
where
header = maybe [] head pattern
footer = maybe [] foot pattern
head pat =
[ "#"
, "# Restricted to " ++ pat
, "#"
]
foot _ =
[ [s|# hidden: %d failures|] (length xs - length ys)
]
xs = filter (failure . snd) (zip [0..] $ M.toAscList outs)
ys = map f $ filter (match . snd) xs
match xs = maybe True (\f -> f `L.isInfixOf` map toLower (show $ snd $ fst xs)) pattern
failure :: (a,(b,Maybe Bool)) -> Bool
failure x
| showSuccess = True
| otherwise = maybe False not $ snd $ snd x
f (n,((m,lbl),(_,_))) = [s| x %s - %s (%d)|] (pretty m) (pretty lbl) n
run_all :: [IO (IO ())] -> IO [ThreadId]
run_all xs = do
ys <- sequence xs
mapM f ys
where
f cmd = forkIO $ cmd
display :: Shared
-> IO (IO ())
display (Shared { .. }) = do
tok <- newEmptyMVar
observe pr_obl tok
observe error_list tok
observe working tok
observe parser_state tok
observe focus tok
observe redraw tok
observe dump_cmd tok
clearScreen
return $ forever $ do
outs <- read_obs pr_obl
es <- read_obs error_list
w <- read_obs working
fil <- read_obs focus
let ys = proof_report fil outs es (w /= 0)
cursorUpLine $ length ys
clearFromCursorToScreenBeginning
forM_ ys $ \x -> do
let lns = lines x
forM_ lns $ \x -> do
putStr x
clearFromCursorToLineEnd
putStrLn ""
let u = M.size $ M.filter (isNothing.snd) outs
st <- read_obs parser_state
du <- isJust `liftM` read_obs dump_cmd
putStr $ [s|number of workers: %d; untried: %d; parser: %s; dumping: %s|] w u (show st) (show du)
clearFromCursorToLineEnd
-- hFlush stdout
putStrLn ""
-- cursorDown 1
-- putStr "-salut-"
threadDelay 500000
takeMVar tok
serialize :: Shared -> IO (IO ())
serialize (Shared { .. }) = do
tok <- newEmptyMVar
observe pr_obl tok
return $ forever $ do
threadDelay 10000000
takeMVar tok
pos <- read_obs pr_obl
let out = pos
-- (pos@(out,_),es) <- takeMVar ser
es <- read_obs error_list
-- dump_pos fname pos
writeFile
(fname ++ ".report")
(unlines $ proof_report' True Nothing out es False)
dump :: Shared -> IO (IO b)
dump (Shared { .. }) = do
tok <- newEmptyMVar
observe dump_cmd tok
return $ forever $ do
takeMVar tok
pat <- read_obs dump_cmd
case pat of
Just pat -> do
pos <- read_obs pr_obl
dump_z3 pat fname pos
write_obs dump_cmd Nothing
Nothing -> return ()
pdfLatex :: Shared -> IO (IO ())
pdfLatex (Shared { .. }) = do
v <- newEmptyMVar
observe system v
observe error_list v
return $ forever $ do
threadDelay 5000000
takeMVar v
readProcess "pdflatex" [fname] ""
summary :: Shared -> IO (IO ())
summary (Shared { .. }) = do
v <- newEmptyMVar
observe system v
return $ forever $ do
threadDelay 10000000
takeMVar v
s <- read_obs system
produce_summaries fname s
keyboard :: Shared -> IO ()
keyboard sh@(Shared { .. }) = do
modify_obs redraw $ return . not
xs <- getLine
let xs' = map toLower xs
ws = words xs'
if xs' == "quit"
then return ()
else do
if xs' == "goto" then do
xs <- read_obs error_list
case xs of
(Error _ (LI fn i _)):_ -> do
open_at i fn
(MLError _ ((_,LI fn i _):|_)):_ -> do
open_at i fn
[] -> return ()
else if xs' == "resetall" then do
modify_obs pr_obl $ \m ->
return $ M.map (\(x,_) -> (x,Nothing)) m
else if xs' == "retry" then do
let f (Just False) = Nothing
f x = x
modify_obs pr_obl $ \m ->
return $ m & traverse._2 %~ f
else if xs' == "unfocus" then do
write_obs focus Nothing
else if take 1 ws == ["dump"]
&& length ws == 2
&& all isDigit (ws ! 1) then do
modify_obs dump_cmd $ \st -> do
if isNothing st then do
pos <- read_obs pr_obl
return $ Just $ Only $ snd $ keys pos ! (read $ ws ! 1)
else return Nothing
else if xs == "dumpfail" then do
modify_obs dump_cmd $ \st -> do
if isNothing st then
return $ Just AllFailed
else return st
else if xs == "dumpall" then do
modify_obs dump_cmd $ \st -> do
if isNothing st then
return $ Just All
else return st
else if take 1 ws == ["focus"] && length ws == 2 then do
write_obs focus $ Just (ws ! 1)
else do
putStrLn $ [s|Invalid command: '%s'|] xs
keyboard sh
run_pipeline :: FilePath -> Params -> IO ()
run_pipeline fname (Params {..}) = do
system <- new_obs empty_system
working <- new_obs 0
error_list <- new_obs []
exit_code <- newEmptyMVar
m <- load_pos fname M.empty
pr_obl <- new_obs m
parser_state <- new_obs (Idle 0)
focus <- new_obs init_focus
dump_cmd <- new_obs Nothing
redraw <- new_obs True
setNumCapabilities 8
let sh = Shared { .. }
ts <- run_all $
[ summary sh
-- , prover sh -- (M.map f m)
, serialize sh
, parser sh
, dump sh
, display sh
, pdfLatex sh
] ++
(guard (not no_verif) >> [prover sh])
keyboard sh
putStrLn "received a 'quit' command"
mapM_ killThread ts
pos <- read_obs pr_obl
dump_pos fname pos
-- return sh
--type Verifier = StablePtr (Shared)
--
----run_verifier :: CString -> IO Verifier
----run_verifier fname = do
---- fname <- peekCString fname
---- sh <- run_pipeline fname
---- newStablePtr sh
--
---- merr <- gets error_msg
---- mpos <- gets failed_po
---- liftIO $ swapMVar mpos $ concatMap g $ toList res
---- g ((x,y),(p,b))
---- | not b = [Ref fname (show y) (1,1)]
---- | otherwise = []
--
--get_error_list :: Verifier -> IO CErrList
--get_error_list v = do
-- Shared { .. } <- deRefStablePtr v
-- err <- read_obs error_list
-- let xs = map f err
-- r <- newIORef (RL [] xs)
-- newStablePtr r
-- where
-- f (Error x (LI fname i j)) = Ref fname x (i,j)
--
--get_proof_obligations :: Verifier -> IO CErrList
--get_proof_obligations v = do
-- Shared { .. } <- deRefStablePtr v
-- pos <- read_obs pr_obl
-- let ys = concatMap (g fname) $ toList pos
-- r <- newIORef (RL [] ys)
-- newStablePtr r
-- where
-- g fname ((_,y),(_,b))
-- | b == Just False = [Ref fname (show y) (1,1)]
-- | otherwise = []
| literate-unitb/literate-unitb | src/Interactive/Pipeline.hs | mit | 14,380 | 1 | 28 | 6,002 | 4,108 | 2,056 | 2,052 | -1 | -1 |
module System.Flannel.CommandSpec
( spec
) where
import System.Flannel.Command
import qualified System.Flannel.Params as P
import Test.Hspec
spec :: Spec
spec = do
describe "runCommand" $ do
it "executes the command" $ do
result <- runCommand P.defaultParams $ isSet "no set"
result `shouldBe` False
describe "isSet" $ do
let params = P.setFlag "good" P.defaultParams
context "when the flag is set" $ do
it "returns True" $ do
result <- runCommand params $ isSet "good"
result `shouldBe` True
context "when the flag is not set" $ do
it "returns False" $ do
result <- runCommand params $ isSet "bad"
result `shouldBe` False
describe "getOption" $ do
let params = P.setOption "good" "alpha" P.defaultParams
context "when the option is set" $ do
it "returns the value" $ do
result <- runCommand params $ getOption "good"
result `shouldBe` Just "alpha"
context "when the option is not set" $ do
it "returns Nothing" $ do
result <- runCommand params $ getOption "bad"
result `shouldBe` Nothing
describe "getArg" $ do
let params = P.setArg "good" "1" P.defaultParams
context "when the arg is set" $ do
it "returns the value" $ do
result <- runCommand params $ getArg "good"
result `shouldBe` Just "1"
context "when the arg is not set" $ do
it "returns Nothing" $ do
result <- runCommand params $ getArg "bad"
result `shouldBe` Nothing
describe "getRemaining" $ do
let params = P.addRemaining ["1", "2"] P.defaultParams
it "returns the remaining arguments" $ do
result <- runCommand params getRemaining
result `shouldBe` ["1", "2"]
describe "run" $ do
it "executes the IO action" $ do
res <- runCommand P.defaultParams . run $ do
fmap (head . lines) $ readFile "LICENSE"
res `shouldBe` "The MIT License (MIT)"
| nahiluhmot/flannel | spec/System/Flannel/CommandSpec.hs | mit | 2,205 | 0 | 20 | 780 | 600 | 274 | 326 | 51 | 1 |
{-# LANGUAGE OverloadedStrings #-}
module PillsHs.Config where
import Clay
-- the maximum width of the container in pixels
pillsMaxWidth :: Size Abs
pillsMaxWidth = px 1024
pillsWideMaxWidth :: Size Abs
pillsWideMaxWidth = px 1180
pillsWiderMaxWidth :: Size Abs
pillsWiderMaxWidth = px 1366
-- the minimum width of the container in pixels, before it switches to a mobile friendly display
pillsMinWidth :: Size Abs
pillsMinWidth = px 599
-- the padding that will be applied to both sides of a column in pixels, also known as gutter
pillsPaddingWidth :: Double
pillsPaddingWidth = 10
| polo2ro/pillshs | src/PillsHs/Config.hs | mit | 588 | 0 | 5 | 98 | 87 | 48 | 39 | 13 | 1 |
module Lambency.UI (
UIWire, WidgetEvent(..), WidgetState(..), Widget(..), screen,
animatedSpriteRenderer, spriteRenderer, colorRenderer,
textRenderer, dynamicTextRenderer,
combineRenderers,
hbox, vbox, glue
) where
--------------------------------------------------------------------------------
import Control.Monad.Reader
import Control.Wire hiding ((.))
import Data.Word
import FRP.Netwire.Input
import qualified Graphics.UI.GLFW as GLFW
import Lambency.Font
import Lambency.GameObject
import Lambency.Sprite
import Lambency.Types
import Linear hiding (trace, identity)
import Prelude hiding (id)
import qualified Yoga as Y
--------------------------------------------------------------------------------
type UIWire a b = GameWire (Y.LayoutInfo, a) b
data WidgetEvent a b
= WidgetEvent'OnMouseOver {
eventLogic :: UIWire a b
}
| WidgetEvent'OnMouseDown {
_eventMouseButton :: GLFW.MouseButton,
eventLogic :: UIWire a b
}
| WidgetEvent'OnKeyDown {
_eventKey :: GLFW.Key,
eventLogic :: UIWire a b
}
data WidgetState a b = WidgetState {
idleLogic :: UIWire a b,
eventHandlers :: [WidgetEvent a b]
}
blankState :: Monoid b => WidgetState a b
blankState = WidgetState (ignoreFst $ mkConst (Right mempty)) []
newtype Widget a b = Widget { getWidgetLayout :: Y.Layout (WidgetState a b) }
widgetRenderFn :: Monoid b =>
TimeStep -> a -> Y.LayoutInfo -> WidgetState a b ->
GameMonad (b, WidgetState a b)
widgetRenderFn dt input lytInfo widgetState =
let eventWire :: WidgetEvent a b -> UIWire a b
eventWire (WidgetEvent'OnMouseDown mb uiw) =
second (mousePressed mb) >>> uiw
eventWire (WidgetEvent'OnKeyDown key uiw) =
second (keyPressed key) >>> uiw
eventWire e@(WidgetEvent'OnMouseOver uiw) = mkGen $ \dt' (lyt, ipt) -> do
(Right (mx, my), _) <- stepWire mouseCursor dt' $ Right undefined
(V2 wx wy) <- windowSize <$> ask
let bx0 = Y.nodeLeft lytInfo / fromIntegral wx
bx1 = (bx0 + Y.nodeWidth lytInfo) / fromIntegral wx
by0 = Y.nodeTop lytInfo / fromIntegral wy
by1 = (by0 + Y.nodeHeight lytInfo) / fromIntegral wy
x = (mx + 1.0) * 0.5
y = (my + 1.0) * 0.5
if x >= bx0 && x <= bx1 && y >= by0 && y <= by1 then
do
(result, uiw') <- stepWire uiw dt' $ Right (lyt, ipt)
return (result, eventWire (WidgetEvent'OnMouseOver uiw'))
else return (Left "Mouse out of bounds", eventWire e)
handleEvent :: WidgetEvent a b -> (Y.LayoutInfo, a) ->
GameMonad (Maybe (b, WidgetEvent a b))
handleEvent event arg = do
(result, uiw') <- stepWire (eventWire event) dt $ Right arg
case result of
Left _ -> return Nothing
Right x -> return $ Just (x, event { eventLogic = uiw' })
handleEvents :: Monoid b => [WidgetEvent a b] -> (Y.LayoutInfo, a) ->
GameMonad (Maybe b, [WidgetEvent a b])
handleEvents events arg =
let eventFn (res, evts) event = do
result <- handleEvent event arg
case result of
Nothing -> return (res, event : evts)
Just (x, e) -> case res of
Nothing -> return (Just x, e : evts)
Just r -> return (Just $ r `mappend` x, e : evts)
in foldM eventFn (Nothing, []) events
wireArg = (lytInfo, input)
in do
(eventResults, events) <- handleEvents (eventHandlers widgetState) wireArg
case eventResults of
Nothing -> do
(result, uiw') <- stepWire (idleLogic widgetState) dt $ Right wireArg
let newState = widgetState { idleLogic = uiw', eventHandlers = events }
case result of
Right x -> return (x, newState)
Left _ -> error "UI wire inhibited?"
Just result -> return (result, widgetState { eventHandlers = events })
ignoreFst :: GameWire b c -> GameWire (a, b) c
ignoreFst logic = mkGen $ \dt (_, ipt) -> do
(result, logic') <- stepWire logic dt $ Right ipt
return (result, ignoreFst logic')
widgetWire :: Monoid b => Widget a b -> GameWire a b
widgetWire (Widget lyt) = mkGen $ \dt input -> do
(result, newLyt) <- Y.foldRender lyt (widgetRenderFn dt input)
return (Right result, widgetWire $ Widget newLyt)
screenPrg :: Monoid b => [Widget a b] -> GameMonad (Widget a b)
screenPrg children = do
(V2 wx wy) <- windowSize <$> ask
return . Widget $
($ blankState) $
Y.withDimensions (fromIntegral wx) (fromIntegral wy) $
Y.vbox (Y.startToEnd $ getWidgetLayout <$> children)
screen :: Monoid b => [Widget a b] -> GameWire a b
screen children = wireFrom (windowSize <$> ask) $ runScreen $
wireFrom (screenPrg children) widgetWire
where
runScreen ui_wire oldWinDims =
let getUIWire False = wireFrom (screenPrg children) widgetWire
getUIWire True = ui_wire
in mkGen $ \dt input -> do
winDims <- windowSize <$> ask
let ui = getUIWire (winDims == oldWinDims)
(result, next_wire') <- stepWire ui dt $ Right input
return (result, runScreen next_wire' winDims)
renderSpriteAt :: Sprite -> Y.LayoutInfo -> GameMonad ()
renderSpriteAt sprite lytInfo = do
let (x, y, w, h) = (
Y.nodeLeft lytInfo,
Y.nodeTop lytInfo,
Y.nodeWidth lytInfo,
Y.nodeHeight lytInfo)
(V2 _ wy) <- windowSize <$> ask
renderUISpriteWithSize sprite (V2 x (fromIntegral wy - y - h)) (V2 w h)
renderStringAt :: Font -> String -> Y.LayoutInfo -> GameMonad()
renderStringAt font str lytInfo = do
let (x, y) = (Y.nodeLeft lytInfo, Y.nodeTop lytInfo)
(V2 _ wy) <- windowSize <$> ask
renderUIString font str $ V2 x (fromIntegral wy - y - stringHeight font str)
animatedRenderer :: Monoid b => GameWire a Sprite -> GameWire a b -> UIWire a b
animatedRenderer spriteWire logic = mkGen $ \dt (lytInfo, val) -> do
(spriteResult, spriteWire') <- stepWire spriteWire dt $ Right val
(logicResult, logic') <- stepWire logic dt $ Right val
case spriteResult of
Right nextSprite -> do
renderSpriteAt nextSprite lytInfo
(nextSpriteResult, _) <- stepWire spriteWire' dt $ Right val
case nextSpriteResult of
Right _ -> return (logicResult, animatedRenderer spriteWire' logic')
Left _ -> return (logicResult, spriteRenderer nextSprite logic')
Left _ -> error "Should never get here"
animatedSpriteRenderer :: Monoid b =>
Sprite -> SpriteAnimationType -> GameWire a b ->
UIWire a b
animatedSpriteRenderer sprite animType logic =
animatedRenderer (animatedWire sprite animType) logic
spriteRenderer :: Monoid b => Sprite -> GameWire a b -> UIWire a b
spriteRenderer s logic = mkGen $ \dt (lytInfo, val) -> do
renderSpriteAt s lytInfo
(result, logic') <- stepWire logic dt $ Right val
return (result, spriteRenderer s logic')
colorRenderer :: Monoid b => V4 Word8 -> GameWire a b -> UIWire a b
colorRenderer color logic = mkGen $ \dt (lytInfo, val) -> do
let byteColor = fromIntegral <$> color
s <- changeSpriteColor byteColor <$> simpleSprite <$> ask
renderSpriteAt s lytInfo
(result, logic') <- stepWire logic dt $ Right val
return (result, spriteRenderer s logic')
textRenderer :: Monoid b => Font -> String -> GameWire a b -> UIWire a b
textRenderer font str logic = mkGen $ \dt (lytInfo, val) -> do
renderStringAt font str lytInfo
(result, logic') <- stepWire logic dt $ Right val
return (result, textRenderer font str logic')
dynamicTextRenderer :: Monoid b => Font -> GameWire a (b, String) -> UIWire a b
dynamicTextRenderer font logic = mkGen $ \dt (lytInfo, val) -> do
(wireResult, logic') <- stepWire logic dt $ Right val
result <- case wireResult of
(Right (bVal, str)) -> do
renderStringAt font str lytInfo
return $ Right bVal
(Left e) -> return (Left e)
return (result, dynamicTextRenderer font logic')
combineRenderers :: Monoid b => [UIWire a b] -> UIWire a b
combineRenderers = mconcat
hbox :: Monoid b => [Widget a b] -> Widget a b
hbox widgets = Widget
$ ($ blankState)
$ Y.stretched
$ Y.hbox (Y.spaceBetween (getWidgetLayout <$> widgets))
vbox :: Monoid b => [Widget a b] -> Widget a b
vbox widgets = Widget
$ ($ blankState)
$ Y.stretched
$ Y.vbox (Y.spaceBetween (getWidgetLayout <$> widgets))
glue :: Monoid b => Widget a b
glue = Widget
$ ($ blankState)
$ Y.growable 2.0 (Y.Min 0.0) (Y.Min 0.0)
$ Y.exact 1.0 1.0
| Mokosha/Lambency | lib/Lambency/UI.hs | mit | 8,644 | 0 | 24 | 2,213 | 3,174 | 1,607 | 1,567 | 186 | 9 |
module DotName where
-- Example from Data.Function
-- uses an infix operator as a pattern variable
-- I think the only solution is to rename to normal variable and make all uses prefix
on :: (b -> b -> c) -> (a -> b) -> a -> a -> c
(.*.) `on` f = \x y -> f x .*. f y
| antalsz/hs-to-coq | examples/tests/DotName.hs | mit | 270 | 0 | 8 | 65 | 78 | 44 | 34 | 3 | 1 |
import Control.Concurrent (threadDelay)
import Control.Exception
import Control.Monad
import qualified Graphics.Rendering.OpenGL as GL
import Graphics.Rendering.OpenGL (($=))
import qualified Graphics.UI.GLFW as GLFW
import Prelude hiding (catch)
main = do
-- initialize has to come first. If it doesn't return True,
-- this crashes with a pattern match error.
True <- GLFW.initialize
-- Set the RGB bits to get a color window.
-- See the GLFW-b docs for all the options
True <- GLFW.openWindow GLFW.defaultDisplayOptions
{ GLFW.displayOptions_numRedBits = 8
, GLFW.displayOptions_numGreenBits = 8
, GLFW.displayOptions_numBlueBits = 8
, GLFW.displayOptions_numDepthBits = 1
, GLFW.displayOptions_width = 640
, GLFW.displayOptions_height = 480
}
GLFW.setWindowSizeCallback $ resize
-- Use `$=` for assigning to GL values, `get` to read them.
-- These functions basically hide IORefs.
GL.depthFunc $= Just GL.Less
-- Use `finally` so that `quit` is called whether or
-- not `mainLoop` throws an exception
finally mainLoop quit
-- | Resize the viewport and set the projection matrix
resize w h = do
-- These are all analogous to the standard OpenGL functions
GL.viewport $= (GL.Position 0 0, GL.Size (fromIntegral w) (fromIntegral h))
GL.matrixMode $= GL.Projection
GL.loadIdentity
GL.perspective 45 (fromIntegral w / fromIntegral h) 1 100
GL.matrixMode $= GL.Modelview 0
-- | Close the window and terminate GLFW
quit = GLFW.closeWindow >> GLFW.terminate
-- | This will print and clear the OpenGL errors
printErrors = GL.get GL.errors >>= mapM_ print
-- | Draw the window and handle input
mainLoop = do
now <- GLFW.getTime
draw now
-- Input is polled each time swapBuffers is called
esc <- GLFW.keyIsPressed GLFW.KeyEsc
isClosed <- fmap not GLFW.windowIsOpen
unless (esc || isClosed) $ do
-- Sleep for the rest of the frame
frameLeft <- fmap (spf + now -) GLFW.getTime
when (frameLeft > 0) $
threadDelay (truncate $ 1000000 * frameLeft)
mainLoop
where
-- maximum frame rate
fps = 60
spf = recip fps
-- | Draw a frame
draw :: Double -> IO ()
draw t = do
-- Again, the functions in GL almost all map to standard OpenGL functions
GL.clear [GL.ColorBuffer, GL.DepthBuffer]
GL.loadIdentity
GL.translate $ GL.Vector3 0 0 (-50 :: GL.GLfloat)
GL.scale 10 10 (1 :: GL.GLfloat)
GL.rotate theta axis
-- renderPrimitive wraps the supplied action with glBegin and glEnd.
-- We'll stop using this when we switch to shaders and vertex buffers.
GL.renderPrimitive GL.Quads $
-- Draw a unit square centered on the origin
forM_ [(0, 0), (1, 0), (1, 1), (0, 1)] $ \(x, y) ->
-- Note that we have to explicitly type Vertex* and Vector*, because
-- they are polymorphic in number field.
let vtx = GL.Vertex3 (x - 0.5) (y - 0.5) 0 :: GL.Vertex3 GL.GLfloat
in GL.vertex vtx
printErrors
GL.flush
GLFW.swapBuffers
where
-- GL.rotate takes the angle in degrees, not radians
theta = realToFrac t * 360
axis = GL.Vector3 0 1 0 :: GL.Vector3 GL.GLfloat
| spetz911/progames | GLFW/glfw01.hs | mit | 3,192 | 1 | 15 | 734 | 744 | 390 | 354 | 55 | 1 |
module HandBrake.T.Encode
( tests )
where
-- base --------------------------------
import Control.Applicative ( pure )
import Data.Function ( ($), (&) )
import Data.List.NonEmpty ( NonEmpty( (:|) ) )
import Data.Tuple ( fst )
import System.Exit ( ExitCode )
import System.IO ( IO )
-- base-unicode-symbols ----------------
import Data.Function.Unicode ( (∘) )
-- fpath -------------------------------
import FPath.AbsDir ( absdir, root )
import FPath.AbsFile ( absfile )
import FPath.PathComponent ( pc )
-- more-unicode ------------------------
import Data.MoreUnicode.Functor ( (⊳) )
import Data.MoreUnicode.Lens ( (⊢), (⊩) )
import Data.MoreUnicode.Maybe ( pattern 𝕵, pattern 𝕹 )
import Data.MoreUnicode.Natural ( ℕ )
import Data.MoreUnicode.String ( 𝕊 )
import Data.MoreUnicode.Text ( 𝕋 )
-- range -------------------------------
import Data.Range ( Bound( Bound ), BoundType( Inclusive )
, Range( LowerBoundRange, SingletonRange ), (+=+) )
-- stdmain -----------------------------
import StdMain.UsageError ( UsageFPathIOError, usageError )
-- tasty -------------------------------
import Test.Tasty ( TestTree, testGroup )
-- tasty-hunit -------------------------
import Test.Tasty.HUnit ( testCase )
-- tasty-plus --------------------------
import TastyPlus ( (≟), assertLeft, assertListEqR
, runTestsP, runTestsReplay, runTestTree )
------------------------------------------------------------
-- local imports --
------------------------------------------------------------
import HandBrake.Encode ( AudioTracks( AudioTracks )
, Chapters( Chapters )
, Numbering( NoNumber, Number, Series )
, Profile( ProfileH265_576P, ProfileH265_720P
, Profile_DeadVideo )
, SubtitleTracks( SubtitleTracks )
, TwoPass( NoTwoPass )
, audios, audioEncoder, chapters, encodeArgs
, encodeRequest, input, name, numbering, outputDir
, outputName, profile, quality, subtitles, titleID
, twoPass
)
--------------------------------------------------------------------------------
tests ∷ TestTree
tests =
testGroup "Encode" $
let
testEncode nm req exp =
testGroup nm $
assertListEqR nm (fst ⊳ encodeArgs @UsageFPathIOError req) exp
base_req = encodeRequest [absfile|/nonesuch|] root 3 (𝕵 "bob")
(AudioTracks $ pure 2)
& subtitles ⊢ (SubtitleTracks [3,4])
usage_error nm txt req =
testCase nm $
assertLeft (usageError @𝕋 @UsageFPathIOError txt ≟) (encodeArgs req)
in
[ testEncode "base_req" base_req
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder", "copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/03-bob.mkv"
]
, testEncode "inputOffset 2" (base_req & numbering ⊢ Number 2)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/05-bob.mkv"
]
, usage_error "inputOffset -3"
"output number 0 (3+(-3)) < 0"
(base_req & numbering ⊢ Number (-3))
, testEncode "NoNumber" (base_req & numbering ⊢ NoNumber)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/bob.mkv"
]
, testEncode "Series S 5" (base_req & numbering ⊢ Series (5,"S") 0)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/S - 05x03 - bob.mkv"
]
, testEncode "Series S 6, no name"
(base_req & numbering ⊢ Series (6,"S") 1 & name ⊢ 𝕹)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/S - 06x04.mkv"
]
, testEncode "chapters 6,7" (base_req & chapters
⊢ (Chapters $ 𝕵 (6 +=+ 7)))
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--chapters", "6-7"
, "--output" , "/03-bob.mkv"
]
, testEncode "chapters 5"
(base_req & chapters ⊢ (Chapters ∘ 𝕵 $ SingletonRange 5))
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--chapters", "5"
, "--output" , "/03-bob.mkv"
]
, testEncode "no two pass" (base_req & twoPass ⊢ NoTwoPass)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/03-bob.mkv"
]
, testEncode "profile 576" (base_req & profile ⊢ ProfileH265_576P)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 576p25"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/03-bob.mkv"
]
, testEncode "audios 8,9" (base_req & audios ⊢ AudioTracks (8 :| [9]) )
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "8,9"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/03-bob.mkv"
]
, testEncode "no subs" (base_req & subtitles ⊢ SubtitleTracks [] )
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--quality" , "26.0"
, "--output" , "/03-bob.mkv"
]
, testEncode "quality 22.5" (base_req & quality ⊢ 22.5 )
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder","copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "22.5"
, "--output" , "/03-bob.mkv"
]
, testEncode "no audio copy" (base_req & audioEncoder ⊢ 𝕵 "mp3")
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder", "mp3"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/03-bob.mkv"
]
, testEncode "no name" (base_req & name ⊢ 𝕹)
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder", "copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/03.mkv"
]
, usage_error "no name, no number"
"no number & no title"
(base_req & name ⊢ 𝕹 & numbering ⊢ NoNumber)
, usage_error "illegal range"
"illegal range «[7-»"
(base_req & chapters ⊢ Chapters (𝕵 $ LowerBoundRange (Bound 7 Inclusive)))
, testEncode "outputDir " (base_req & outputDir ⊢ [absdir|/out/|])
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder", "copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/out/03-bob.mkv"
]
, testEncode "outputName " (base_req & outputName ⊩ [pc|output.mkv|])
[ "--input" , "/nonesuch"
, "--title" , "3"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--two-pass", "--turbo"
, "--preset", "H.265 MKV 2160p60"
, "--aencoder", "copy"
, "--audio" , "2"
, "--subtitle", "3,4", "--subtitle-default", "0"
, "--quality" , "26.0"
, "--output" , "/output.mkv"
]
, testEncode "altogether now"
(base_req & input ⊢ [absfile|/not-here|]
& titleID ⊢ 5
& numbering ⊢ Series (7,"T") 1
& name ⊢ 𝕹
& chapters ⊢ Chapters (𝕵 $ 7 +=+ 9)
& twoPass ⊢ NoTwoPass
& profile ⊢ ProfileH265_720P
& audios ⊢ AudioTracks (2 :| [1])
& subtitles ⊢ SubtitleTracks []
& quality ⊢ 26
& audioEncoder ⊢ 𝕵 "flac24,av_aac"
& outputDir ⊢ [absdir|/outdir/|]
& outputName ⊩ [pc|out.mkv|])
[ "--input" , "/not-here"
, "--title" , "5"
, "--markers"
, "--deinterlace"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--preset", "H.265 MKV 720p30"
, "--aencoder", "flac24,av_aac"
, "--audio" , "2,1"
, "--quality" , "26.0"
, "--chapters", "7-9"
, "--output" , "/outdir/out.mkv"
]
, testEncode "dead video"
(base_req & input ⊢ [absfile|/not-here|]
& titleID ⊢ 5
& numbering ⊢ Series (7,"T") 1
& name ⊢ 𝕹
& chapters ⊢ Chapters (𝕵 $ 7 +=+ 9)
& profile ⊢ Profile_DeadVideo
& audios ⊢ AudioTracks (2 :| [1])
& subtitles ⊢ SubtitleTracks []
& quality ⊢ 26
& audioEncoder ⊢ 𝕵 "flac24,av_aac"
& outputDir ⊢ [absdir|/outdir/|]
& outputName ⊩ [pc|out.mkv|])
[ "--input" , "/not-here"
, "--title" , "5"
, "--markers"
, "--audio-copy-mask", "aac,ac3,eac3,truehd,dts,dtshd,mp3,flac"
, "--preset", "H.265 MKV 480p30"
, "--aencoder", "flac24,av_aac"
, "--audio" , "2,1"
, "--quality" , "26.0"
, "--chapters", "7-9"
, "--output" , "/outdir/out.mkv"
]
]
----------------------------------------
_test ∷ IO ExitCode
_test = runTestTree tests
--------------------
_tests ∷ 𝕊 → IO ExitCode
_tests = runTestsP tests
_testr ∷ 𝕊 → ℕ → IO ExitCode
_testr = runTestsReplay tests
-- that's all, folks! ----------------------------------------------------------
| sixears/handbrake | src/HandBrake/T/Encode.hs | mit | 16,599 | 43 | 35 | 7,017 | 2,751 | 1,697 | 1,054 | -1 | -1 |
import Diagrams.Prelude
import Diagrams.Backend.Cairo
import Data.Default
import Control.Lens ((^.))
import Data.List
import Data.Ord
import Diagrams.Plots
-- | project a 3d point to 2d
projection :: (Double, Double, Double) -- ^ position of camera
-> (Double, Double, Double) -- ^ orientation of camera
-> (Double, Double, Double)
-> (Double, Double, Double) -- ^ 3d point to be projected
-> (Double, Double)
projection (cx',cy',cz') (θx,θy,θz) (ex,ey,ez) (ax,ay,az) = (bx, by)
where
bx = ez / dz * dx - ex
by = ez / dz * dy - ey
dx = cy * (sz * y + cz * x) - sy * z
dy = sx * (cy * z + sy * (sz * y + cz * x)) + cx * (cz * y - sz * x)
dz = cx * (cy * z + sy * (sz * y + cz * x)) - sx * (cz * y - sz * x)
x = ax - cx'
y = ay - cy'
z = az - cz'
sx = sin θx
sy = sin θy
sz = sin θz
cx = cos θx
cy = cos θy
cz = cos θz
ps :: [(Double, Double, Double)]
ps = [(2,2,2), (2,5,2), (5,5,2), (5,2,2)]
ps' :: [(Double, Double, Double)]
ps' = [(3,3,1), (3,4,1), (4,4,1), (4,3,1)]
nodes = zipWith f [0..] $ map (projection (0,0,0) (0,0,0) (0,0,1)) $ ps ++ ps'
where
f i (x,y) = def {_nodeId = i, _nodeX = x, _nodeY = y}
{-
nodes = [ def { _nodeId=0, _nodeX=2, _nodeY=2, _nodeSize=0.1, _nodeOpacity=0.5, _nodeColor=red }
, def { _nodeId=1, _nodeX=2, _nodeY=5, _nodeSize=0.03 }
, def { _nodeId=2, _nodeX=5, _nodeY=5, _nodeColor=green }
, def { _nodeId=3, _nodeX=5, _nodeY=2, _nodeColor=green }
]
-}
edges = [ def { _edgeId = 0, _edgeFrom = 0, _edgeTo = 1 }
, def { _edgeId = 1, _edgeFrom = 1, _edgeTo = 2 }
, def { _edgeId = 3, _edgeFrom = 2, _edgeTo = 3 }
, def { _edgeId = 4, _edgeFrom = 3, _edgeTo = 0 }
, def { _edgeId = 5, _edgeFrom = 4, _edgeTo = 5 }
, def { _edgeId = 6, _edgeFrom = 5, _edgeTo = 6 }
, def { _edgeId = 7, _edgeFrom = 6, _edgeTo = 7 }
, def { _edgeId = 8, _edgeFrom = 7, _edgeTo = 4 }
]
main = do
let xaxis = realAxis (xlo,xhi) 0.2 def
yaxis = realAxis (ylo,yhi) 0.2 def
area = plotArea 5.5 4.8 (yaxis, emptyAxis, emptyAxis, xaxis)
plot = drawGraph (nodes, edges)
p = area <+ (plot, BL)
xlo = _nodeX $ minimumBy (comparing _nodeX) nodes
xhi = _nodeX $ maximumBy (comparing _nodeX) nodes
ylo = _nodeY $ minimumBy (comparing _nodeY) nodes
yhi = _nodeY $ maximumBy (comparing _nodeY) nodes
renderCairo "1.png" (Dims 480 480) $ showPlot p
| kaizhang/haskell-plot | Test.hs | mit | 2,558 | 0 | 13 | 775 | 1,082 | 629 | 453 | 52 | 1 |
{-# LANGUAGE CPP, FlexibleContexts, JavaScriptFFI, OverloadedStrings #-}
import Control.Concurrent
import Control.Concurrent.STM
import Control.Exception
import Control.Monad
import Control.Monad.State
import Data.IORef
import qualified Data.Serialize as S
import qualified Data.Text as T
import System.Random
import Flaw.App
import Flaw.Asset
import Flaw.Asset.FolderAssetPack
import Flaw.Asset.RemapAssetPack
import Flaw.Book
import Flaw.Graphics
import Flaw.Graphics.Program
import Flaw.Graphics.Sampler
import Flaw.Math
import Flaw.Math.Geometry
import Flaw.Input
import Flaw.Input.Mouse
import Flaw.Input.Keyboard
import Flaw.Visual.Geometry
import Flaw.Window
#if defined(ghcjs_HOST_OS)
import GHCJS.Types
import GHCJS.Foreign.Callback
import GHCJS.Marshal.Pure
#endif
data GameState = GameState
{ gsPhase :: GamePhase
, gsCameraAlpha :: Float
, gsCameraBeta :: Float
, gsCameraDistance :: Float
, gsLightAngle :: Float
, gsActors :: [Actor]
, gsFirstCursor :: Maybe ((Int, Int), (Int, Int))
, gsUserActorType :: ActorType
, gsUserGun :: GunState
, gsComputerGun :: GunState
, gsDamages :: [Damage]
, gsBeaverLives :: Int
, gsPekaLives :: Int
, gsUserSpawn :: Maybe Float2
} deriving Show
data GamePhase = GameBattle | GameFinish deriving (Eq, Show)
data GunState = GunState
{ gunStateTime :: Float
} deriving Show
data Actor = Actor
{ actorType :: ActorType
, actorStartPosition :: !Float2
, actorFinishPosition :: !Float2
, actorTime :: Float
, actorTotalTime :: Float
, actorState :: ActorState
, actorAngle :: Float
} deriving Show
data ActorType = Peka | Beaver deriving (Eq, Show)
actorFlySpeed :: Float
actorFlySpeed = 11.3
actorGroundSpeed :: Float
actorGroundSpeed = 1.27
actorAngleSpeed :: Float
actorAngleSpeed = actorGroundSpeed / actorOffset
gravity :: Float
gravity = 22.5
actorOffset :: Float
actorOffset = 0.127
gunCoolDown :: Float
gunCoolDown = 0.5
actorDeadTime :: Float
actorDeadTime = 5
actorExplodeTime :: Float
actorExplodeTime = 0.5
actorExplodeDistance :: Float
actorExplodeDistance = 0.127
actorWinningOffset :: Float
actorWinningOffset = 1.27
actorWinningScale :: Float
actorWinningScale = 5
actorWinningTime :: Float
actorWinningTime = 1
livesAmount :: Int
livesAmount = 50
moveClickThreshold :: Num a => a
moveClickThreshold = 50
data ActorState = ActorFlying Float | ActorRunning | ActorDead | ActorExplode | ActorWinning deriving (Eq, Show)
calcActorPosition :: Actor -> Float3
calcActorPosition Actor
{ actorStartPosition = Vec2 sx sy
, actorFinishPosition = Vec2 fx fy
, actorTime = t
, actorTotalTime = tt
, actorState = as
} = case as of
ActorFlying angle -> Vec3 (sx * (1 - k) + fx * k) (sy * (1 - k) + fy * k) z where
k = t / tt
z = actorOffset + actorFlySpeed * (sin angle) * t - gravity * t * t / 2
ActorRunning -> Vec3 (sx * (1 - k) + fx * k) (sy * (1 - k) + fy * k) actorOffset where
k = t / tt
ActorDead -> Vec3 sx sy actorOffset
ActorExplode -> Vec3 (sx * (1 - k) + fx * k) (sy * (1 - k) + fy * k) actorOffset where
k = t / tt
ActorWinning -> Vec3 sx sy actorOffset
spawnActor :: ActorType -> Float2 -> Float2 -> Maybe Actor
spawnActor at s f = maybeActor where
sin2angle = (norm $ s - f) * gravity / (actorFlySpeed * actorFlySpeed)
angle = 0.5 * (pi - asin sin2angle)
maybeActor = if sin2angle >= 1 || norm (castlePosition at - f) < 1.27 then Nothing else Just Actor
{ actorType = at
, actorStartPosition = s
, actorFinishPosition = f
, actorTime = 0
, actorTotalTime = 2 * actorFlySpeed * (sin angle) / gravity
, actorState = ActorFlying angle
, actorAngle = 0
}
castlePosition :: ActorType -> Float2
castlePosition at = case at of
Peka -> Vec2 0 5
Beaver -> Vec2 0 (-5)
castleLine :: ActorType -> Float
castleLine at = case at of
Peka -> 3.8
Beaver -> -3.8
fieldWidth :: Float
fieldWidth = 2.5
enemyActor :: ActorType -> ActorType
enemyActor at = case at of
Peka -> Beaver
Beaver -> Peka
data Damage = Damage ActorType Float2 deriving Show
initialGameState :: GameState
initialGameState = GameState
{ gsPhase = GameBattle
, gsCameraAlpha = 0
, gsCameraBeta = 0.35
, gsCameraDistance = 10
, gsLightAngle = 0
, gsActors = []
, gsFirstCursor = Nothing
, gsUserActorType = Peka
, gsUserGun = GunState
{ gunStateTime = 0
}
, gsComputerGun = GunState
{ gunStateTime = 0
}
, gsDamages = []
, gsBeaverLives = livesAmount
, gsPekaLives = livesAmount
, gsUserSpawn = Nothing
}
getFrontScreenPoint :: (Vectorized a, Fractional a) => Mat4x4 a -> Vec3 a -> Vec3 a
getFrontScreenPoint (Mat4x4
m11 m12 m13 m14
m21 m22 m23 m24
m31 m32 m33 m34
m41 m42 m43 m44
) (Vec3 sx sy sz) = Vec3 (dx / d) (dy / d) (dz / d) where
a11 = m11 - sx * m41
a12 = m12 - sx * m42
a13 = m13 - sx * m43
a14 = sx * m44 - m14
a21 = m21 - sy * m41
a22 = m22 - sy * m42
a23 = m23 - sy * m43
a24 = sy * m44 - m24
a31 = m31 - sz * m41
a32 = m32 - sz * m42
a33 = m33 - sz * m43
a34 = sz * m44 - m34
d = a11 * (a22 * a33 - a23 * a32) - a12 * (a21 * a33 - a23 * a31) + a13 * (a21 * a32 - a22 * a31)
dx = a14 * (a22 * a33 - a23 * a32) - a12 * (a24 * a33 - a23 * a34) + a13 * (a24 * a32 - a22 * a34)
dy = a11 * (a24 * a33 - a23 * a34) - a14 * (a21 * a33 - a23 * a31) + a13 * (a21 * a34 - a24 * a31)
dz = a11 * (a22 * a34 - a24 * a32) - a12 * (a21 * a34 - a24 * a31) + a14 * (a21 * a32 - a22 * a31)
intersectRay :: (Vectorized a, Fractional a) => Vec3 a -> Vec3 a -> Vec3 a -> a -> Vec3 a
intersectRay a d n nq = a + d * vecFromScalar ((nq - dot a n) / (dot d n))
affineActorLookAt :: (Vectorized a, Floating a) => Vec3 a -> Vec3 a -> Vec3 a -> Mat4x4 a
affineActorLookAt position@(Vec3 px py pz) target direction = r where
y@(Vec3 yx yy yz) = normalize $ target - position
x@(Vec3 xx xy xz) = normalize $ cross y direction
Vec3 zx zy zz = cross x y
r = Mat4x4
xx yx zx px
xy yy zy py
xz yz zz pz
0 0 0 1
main :: IO ()
main = handle errorHandler $ withApp appConfig
{ appConfigTitle = "PEKABEAVER"
, appConfigNeedDepthBuffer = True
} $ \window device context presenter inputManager -> withBook $ \bk -> do
setWindowMouseCursor window MouseCursorHand
--setWindowMouseLock window True
-- run detection of closed window
windowLoopVar <- newEmptyMVar
windowEventsChan <- atomically $ chanWindowEvents window
_ <- forkIO $ do
let loop = do
event <- atomically $ readTChan windowEventsChan
case event of
DestroyWindowEvent -> putMVar windowLoopVar True
_ -> loop
loop
-- run input processing thread
keyboardChan <- atomically $ chanInputEvents inputManager
mouseChan <- atomically $ chanInputEvents inputManager
-- initial input states
keyboardState <- atomically initialInputState
mouseState <- atomically initialInputState
-- load asset pack
assetPack <- do
Right assetPackContainer <- S.decode <$> loadAsset (FolderAssetPack "") "pack.bin"
return $ loadRemapAssetPack assetPackContainer (FolderAssetPack "assetpack/") :: IO (RemapAssetPack FolderAssetPack T.Text)
let loadTextureAsset = createNativeTexture
-- load field
Geometry
{ geometryVertexBuffer = vbField
, geometryIndexBuffer = ibField
, geometryIndicesCount = icField
} <- book bk (loadGeometryAsset device =<< loadAsset assetPack "field.bin")
tField <- book bk $ loadTextureAsset device defaultSamplerStateInfo =<< loadAsset assetPack "castle.jpg"
-- load beaver
Geometry
{ geometryVertexBuffer = vbBeaver
, geometryIndexBuffer = ibBeaver
, geometryIndicesCount = icBeaver
} <- book bk (loadGeometryAsset device =<< loadAsset assetPack "beaver.bin")
tBeaver <- book bk $ loadTextureAsset device defaultSamplerStateInfo =<< loadAsset assetPack "beaver.jpg"
-- load peka
Geometry
{ geometryVertexBuffer = vbPeka
, geometryIndexBuffer = ibPeka
, geometryIndicesCount = icPeka
} <- book bk (loadGeometryAsset device =<< loadAsset assetPack "peka.bin")
tPeka <- book bk $ loadTextureAsset device defaultSamplerStateInfo =<< loadAsset assetPack "peka.png"
let samplerState = nullSamplerState
-- program
ubsCamera <- uniformBufferSlot 0
uViewProj <- uniform ubsCamera
uCameraPosition <- uniform ubsCamera
ubsLight <- uniformBufferSlot 1
uLightPosition <- uniform ubsLight
--ubsMaterial <- uniformBufferSlot 2
--uDiffuseColor <- uniform ubsMaterial
ubsObject <- uniformBufferSlot 3
uWorld <- uniform ubsObject
usCamera <- book bk $ createUniformStorage device ubsCamera
usLight <- book bk $ createUniformStorage device ubsLight
--usMaterial <- book bk $ createUniformStorage device ubsMaterial
usObject <- book bk $ createUniformStorage device ubsObject
program <- book bk $ createProgram device $ do
aPosition <- attribute 0 0 0 (AttributeVec3 AttributeFloat32)
aNormal <- attribute 0 12 0 (AttributeVec3 AttributeFloat32)
aTexcoord <- attribute 0 24 0 (AttributeVec2 AttributeFloat32)
worldPosition <- temp $ mul uWorld $ cvec31 aPosition (constf 1)
worldNormal <- temp $ mul uWorld $ cvec31 aNormal (constf 0)
rasterize (mul uViewProj worldPosition) $ do
let toLight = normalize $ (xyz__ worldPosition) - uLightPosition
--diffuse <- temp $ max_ 0 $ dot toLight $ xyz__ worldNormal
diffuse <- temp $ min_ (constf 1) $ constf 0.5 + (abs $ dot toLight $ normalize $ xyz__ worldNormal)
diffuseColor <- temp $ sample (sampler2D3f 0) aTexcoord
colorTarget 0 $ cvec31 (diffuseColor * vecFromScalar diffuse) (constf 1)
let
gameStep :: Float -> StateT GameState IO ()
gameStep frameTime = do
-- check exit
#if !defined(ghcjs_HOST_OS)
loop <- liftIO $ tryTakeMVar windowLoopVar
case loop of
Just True -> liftIO $ exitApp
_ -> return ()
#endif
cameraPosition <- do
s <- get
let alpha = gsCameraAlpha s
let beta = gsCameraBeta s
let distance = gsCameraDistance s
return $ Vec3 (distance * (cos alpha * cos beta)) (distance * (sin alpha * cos beta)) (distance * sin beta)
rs <- get
(viewProj, viewportWidth, viewportHeight) <- liftIO $ render context $ do
present presenter $ do
renderClearColor 0 (Vec4 0.5 0.5 0.5 1)
renderClearDepth 0
renderDepthTestFunc DepthTestFuncGreater
renderProgram program
Vec4 viewportLeft viewportTop viewportRight viewportBottom <- renderGetViewport
let viewportWidth = viewportRight - viewportLeft
let viewportHeight = viewportBottom - viewportTop
let aspect = (fromIntegral viewportWidth) / (fromIntegral viewportHeight)
let view = affineLookAt cameraPosition (Vec3 0 0 0) (Vec3 0 0 1)
let proj = projectionPerspectiveFov (pi / 4) aspect (-50 :: Float) (-0.01)
let viewProj = mul proj view
renderUniform usCamera uViewProj viewProj
renderUniform usCamera uCameraPosition cameraPosition
renderUploadUniformStorage usCamera
renderUniformStorage usCamera
renderUniform usLight uLightPosition $ let
angle = gsLightAngle rs
in Vec3 (2 * cos angle) (2 * sin angle) 2
renderUploadUniformStorage usLight
renderUniformStorage usLight
--renderUniform usMaterial uDiffuseColor $ Vec3 1 0 0
--renderUploadUniformStorage usMaterial
--renderUniformStorage usMaterial
-- render field
renderUniform usObject uWorld $ affineTranslation ((Vec3 0 0 0) :: Float3)
renderUploadUniformStorage usObject
renderUniformStorage usObject
renderVertexBuffer 0 vbField
renderIndexBuffer ibField
renderSampler 0 tField samplerState
renderDraw icField
-- render actors
forM_ (gsActors rs) $ \actor@Actor
{ actorType = at
, actorState = as
, actorFinishPosition = Vec2 fx fy
, actorTime = t
, actorTotalTime = tt
, actorAngle = aa
} -> do
let (vb, ib, ic, tex) = case at of
Peka -> (vbPeka, ibPeka, icPeka, tPeka)
Beaver -> (vbBeaver, ibBeaver, icBeaver, tBeaver)
let k = t / tt
let position = calcActorPosition actor
let translation = affineActorLookAt position (Vec3 fx fy actorOffset) (Vec3 0 0 1)
let world = case as of
ActorFlying _ -> mul translation $ affineFromQuat $ affineAxisRotation (Vec3 (-1) 0 0) $ k * pi * 2
ActorRunning -> if at == Peka then mul translation $ affineFromQuat $ affineAxisRotation (Vec3 (-1) 0 0) aa else translation
ActorDead -> mul translation $ mul (affineTranslation $ Vec3 0 0 $ 0.05 - actorOffset) $ affineScaling (Vec3 1.5 1.5 (0.1 :: Float))
ActorExplode ->
--mul translation $ mul (affineTranslation $ Vec3 0 0 $ k * 10) $ affineScaling $ vecFromScalar $ 1 + k * 0.5
mul translation $ affineScaling $ Vec3 1 (1 * (1 - k) + 0.1 * k) 1
ActorWinning -> mul translation $ mul (affineTranslation $ Vec3 0 0 $ k * actorWinningOffset) $ affineScaling $ vecFromScalar $ 1 - k + actorWinningScale * k
renderUniform usObject uWorld world
renderUploadUniformStorage usObject
renderUniformStorage usObject
renderVertexBuffer 0 vb
renderIndexBuffer ib
renderSampler 0 tex samplerState
renderDraw ic
return (viewProj, viewportWidth, viewportHeight)
-- process input
do
let getMousePoint = do
(cursorX, cursorY) <- liftIO $ atomically $ getMouseCursor mouseState
let frontPoint = getFrontScreenPoint viewProj $ Vec3
((fromIntegral cursorX) / (fromIntegral viewportWidth) * 2 - 1)
(1 - (fromIntegral cursorY) / (fromIntegral viewportHeight) * 2)
0
return $ intersectRay cameraPosition (normalize (frontPoint - cameraPosition)) (Vec3 0 0 1) 0
let
processKeyboard = readTChan keyboardChan >>= \keyboardEvent -> return $ do
case keyboardEvent of
KeyDownEvent KeyEscape -> liftIO $ putMVar windowLoopVar True
_ -> return ()
liftIO $ atomically $ applyInputEvent keyboardState keyboardEvent
processEvent
processMouse = readTChan mouseChan >>= \mouseEvent -> return $ do
case mouseEvent of
MouseDownEvent LeftMouseButton -> do
cursor <- liftIO $ atomically $ getMouseCursor mouseState
state $ \s -> ((), s
{ gsFirstCursor = Just (cursor, cursor)
})
MouseUpEvent LeftMouseButton -> do
(cursorX, cursorY) <- liftIO $ atomically $ getMouseCursor mouseState
s1 <- get
case gsFirstCursor s1 of
Just ((firstCursorX, firstCursorY), _) -> do
if (abs $ cursorX - firstCursorX) < moveClickThreshold && (abs $ cursorY - firstCursorY) < moveClickThreshold then do
(Vec3 fx fy _) <- getMousePoint
state $ \s -> ((), s
{ gsUserSpawn = Just $ Vec2 fx fy
})
else return ()
state $ \s -> ((), s
{ gsFirstCursor = Nothing
})
Nothing -> return ()
CursorMoveEvent cursorX cursorY -> do
s <- get
case gsFirstCursor s of
Just (firstCursor@(firstCursorX, firstCursorY), (moveCursorX, moveCursorY)) -> do
if (abs $ cursorX - firstCursorX) >= moveClickThreshold || (abs $ cursorY - firstCursorY) >= moveClickThreshold then do
put $ s
{ gsCameraAlpha = gsCameraAlpha s - (fromIntegral $ cursorX - moveCursorX) * 0.005
, gsCameraBeta = gsCameraBeta s + (fromIntegral $ cursorY - moveCursorY) * 0.01
, gsFirstCursor = Just (firstCursor, (cursorX, cursorY))
}
else
put $ s
{ gsFirstCursor = Just (firstCursor, (cursorX, cursorY))
}
Nothing -> return ()
RawMouseMoveEvent _dx _dy dz -> state $ \s -> ((), s
{ gsCameraDistance = max 2.5 $ min 12.7 $ dz * (-0.0025) + gsCameraDistance s
})
_ -> return ()
liftIO $ atomically $ applyInputEvent mouseState mouseEvent
processEvent
processEvent = do
action <- liftIO $ atomically $ orElse (orElse processKeyboard processMouse) (return $ return ())
action
processEvent
-- process camera rotation
do
up <- liftIO $ atomically $ getKeyState keyboardState KeyUp
down <- liftIO $ atomically $ getKeyState keyboardState KeyDown
left <- liftIO $ atomically $ getKeyState keyboardState KeyLeft
right <- liftIO $ atomically $ getKeyState keyboardState KeyRight
state $ \s -> ((), s
{ gsCameraAlpha = gsCameraAlpha s + ((if right then 1 else 0) - (if left then 1 else 0)) * frameTime
, gsCameraBeta = max 0.1 $ min 1.5 $ gsCameraBeta s + ((if up then 1 else 0) - (if down then 1 else 0)) * frameTime
})
-- step actors
let stepActor actor@Actor
{ actorState = as
, actorFinishPosition = f
, actorType = at
, actorTime = t
, actorTotalTime = tt
} = do
case as of
ActorFlying _ -> if t >= tt then do
let finishPosition = Vec2 (x_ f) $ castleLine $ enemyActor at
state $ \s -> ((), s
{ gsDamages = (Damage at f) : gsDamages s
})
return [actor
{ actorTime = 0
, actorTotalTime = norm (finishPosition - f) / actorGroundSpeed
, actorStartPosition = f
, actorFinishPosition = finishPosition
, actorState = ActorRunning
, actorAngle = 0
}]
else return [actor
{ actorTime = t + frameTime
}]
ActorRunning -> do
if t >= tt then do
let finishPosition = castlePosition $ enemyActor at
let len = norm $ finishPosition - f
if len < 0.25 then do
state $ \s -> ((), case at of
Beaver -> s { gsPekaLives = gsPekaLives s - 1 }
Peka -> s { gsBeaverLives = gsBeaverLives s - 1 }
)
return [actor
{ actorTime = 0
, actorTotalTime = actorWinningTime
, actorState = ActorWinning
, actorStartPosition = finishPosition
, actorFinishPosition = finishPosition
}]
else return [actor
{ actorTime = 0
, actorTotalTime = len / actorGroundSpeed
, actorStartPosition = f
, actorFinishPosition = finishPosition
, actorState = ActorRunning
, actorAngle = 0
}]
else return [actor
{ actorTime = t + frameTime
, actorAngle = actorAngle actor + actorAngleSpeed * frameTime
}]
ActorDead -> do
if t >= tt then return []
else return [actor
{ actorTime = t + frameTime
}]
ActorExplode -> do
if t >= tt then return []
else return [actor
{ actorTime = t + frameTime
}]
ActorWinning -> do
if t >= tt then return []
else return [actor
{ actorTime = t + frameTime
}]
do
s1 <- get
newActors <- liftM concat $ mapM stepActor $ gsActors s1
state $ \s -> ((), s
{ gsActors = newActors
})
-- apply damages
do
s <- get
let applyDamages (Damage eat ep) actors = map (\actor@Actor
{ actorType = at
, actorState = as
} -> let
Vec3 px py _pz = calcActorPosition actor
in
if at == eat || as == ActorDead || norm (ep - (Vec2 px py)) > 0.5 then actor
else actor
{ actorState = ActorDead
, actorTime = 0
, actorTotalTime = actorDeadTime
, actorStartPosition = Vec2 px py
}) actors
put s
{ gsActors = foldr applyDamages (gsActors s) $ gsDamages s
, gsDamages = []
}
-- annigilate ground actors
state $ \s -> let actors = gsActors s in ((), s
{ gsActors = map (\actor -> let
p@(Vec3 px py _pz) = calcActorPosition actor
at = actorType actor
keep = actorState actor /= ActorRunning || all (\actor2 -> let
p2 = calcActorPosition actor2
at2 = actorType actor2
as2 = actorState actor2
in at == at2 || as2 /= ActorRunning || norm (p - p2) > 0.25
) actors
in if keep then actor
else let startPosition = Vec2 px py in actor
{ actorState = ActorExplode
, actorTime = 0
, actorTotalTime = actorExplodeTime
, actorStartPosition = startPosition
, actorFinishPosition = startPosition + normalize (actorFinishPosition actor - startPosition) * vecFromScalar actorExplodeDistance
}
) actors
})
-- process user's gun
do
s1 <- get
if gsPhase s1 == GameBattle && gunStateTime (gsUserGun s1) <= 0 then do
case gsUserSpawn s1 of
Just position -> do
let at = gsUserActorType s1
case spawnActor at (castlePosition at) position of
Just actor -> state $ \s -> ((), s
{ gsActors = actor : gsActors s
, gsUserGun = (gsUserGun s)
{ gunStateTime = gunCoolDown
}
})
Nothing -> return ()
state $ \s -> ((), s
{ gsUserSpawn = Nothing
})
Nothing -> return ()
else return ()
-- process computer's gun
do
s <- get
if gsPhase s == GameBattle && gunStateTime (gsComputerGun s) <= 0 then do
let minx = -fieldWidth
let maxx = fieldWidth
let at = enemyActor $ gsUserActorType s
let cl = castleLine at
let miny = if cl > 0 then 0 else cl
let maxy = if cl > 0 then cl else 0
x <- liftIO $ getStdRandom $ randomR (minx, maxx)
y <- liftIO $ getStdRandom $ randomR (miny, maxy)
case spawnActor at (castlePosition at) (Vec2 x y) of
Just actor -> put s
{ gsActors = actor : gsActors s
, gsComputerGun = (gsComputerGun s)
{ gunStateTime = gunCoolDown
}
}
Nothing -> return ()
else return ()
-- process gun cooldowns
let processGun gs = gs { gunStateTime = gunStateTime gs - frameTime }
state $ \s -> ((), s
{ gsUserGun = processGun $ gsUserGun s
, gsComputerGun = processGun $ gsComputerGun s
})
-- step light
state $ \s -> ((), s
{ gsLightAngle = gsLightAngle s + frameTime * 3
})
#if defined(ghcjs_HOST_OS)
-- update lives
get >>= \s -> liftIO $ do
if gsPhase s == GameBattle then do
js_setStyleWidth (pToJSVal $ T.pack "beaver_lives") $ pToJSVal $ T.pack $ show $ (fromIntegral $ gsBeaverLives s) * (100 :: Float) / fromIntegral livesAmount
js_setStyleWidth (pToJSVal $ T.pack "peka_lives") $ pToJSVal $ T.pack $ show $ (fromIntegral $ gsPekaLives s) * (100 :: Float) / fromIntegral livesAmount
else return ()
-- check end
get >>= \s -> do
if gsPhase s == GameBattle then do
let beaverLives = gsBeaverLives s
let pekaLives = gsPekaLives s
if beaverLives <= 0 || pekaLives <= 0 then do
put s { gsPhase = GameFinish }
let beaverWon = beaverLives > 0
let userWon = beaverWon == (gsUserActorType s == Beaver)
liftIO $ js_end (pToJSVal $ T.pack $ if beaverWon then "beaver" else "peka") $ pToJSVal $ T.pack $ if userWon then "You won!" else "You lose!"
else return ()
else return ()
-- register start functions
gameStateVar <- liftIO $ newEmptyMVar
do
let start at alpha = do
js_start
putMVar gameStateVar initialGameState
{ gsUserActorType = at
, gsCameraAlpha = alpha
}
beaverStart <- syncCallback ThrowWouldBlock $ start Beaver $ (-pi) / 2
pekaStart <- syncCallback ThrowWouldBlock $ start Peka $ pi / 2
js_registerStart beaverStart pekaStart
-- main loop
runApp $ \frameTime -> modifyMVar_ gameStateVar $ execStateT (gameStep frameTime)
foreign import javascript unsafe "document.getElementById($1).style.width=$2+'%'" js_setStyleWidth :: JSVal -> JSVal -> IO ()
foreign import javascript unsafe "document.getElementById('start-beaver').addEventListener('click', $1, false);document.getElementById('start-peka').addEventListener('click', $2, false);" js_registerStart :: Callback (IO ()) -> Callback (IO ()) -> IO ()
foreign import javascript unsafe "document.getElementById('start').style.display='none';" js_start :: IO ()
foreign import javascript unsafe "document.getElementById('end-'+$1).style.display='block'; document.getElementById('end-title').innerText=$2; document.getElementById('end').style.display='block';" js_end :: JSVal -> JSVal -> IO ()
#else
-- main loop
do
gameStateRef <- newIORef $ initialGameState
{ gsUserActorType = Peka
, gsCameraAlpha = pi / 2
}
runApp $ \frameTime -> writeIORef gameStateRef =<< execStateT (gameStep frameTime) =<< readIORef gameStateRef
#endif
errorHandler :: SomeException -> IO ()
errorHandler = putStrLn . show
| quyse/pekabeaver | Main.hs | mit | 24,202 | 341 | 34 | 5,982 | 8,019 | 4,120 | 3,899 | 527 | 42 |
doubleMe x = x + x
doubleUs x y = x*2 + y*2
lucky :: (Integral a) => a -> String
lucky 7 = "LUCKY NUMBER SEVEN!"
lucky x = "Sorry, you're out of luck, pal!"
maximum' :: (Ord a) => [a] -> a
maximum' [] = error "maximum of empty list"
maximum' [x] = x
maximum' (x:xs)
| x > maxTail = x
| otherwise = maxTail
where maxTail = maximum' xs
data Trafficlight = Red | Yellow | Green deriving (Show)
instance Eq Trafficlight where
Red == Red = True
Green == Green = True
Yellow == Yellow = True
_ == _ = False
| dalonng/hellos | haskell.hello/learnyouahaskell01.hs | gpl-2.0 | 549 | 0 | 8 | 155 | 234 | 117 | 117 | 18 | 1 |
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE FlexibleInstances #-}
module Lambda where
import Data.List (elemIndex, (\\), genericIndex)
import Data.Set (Set, toList, delete, union, singleton)
import Control.Arrow
data NamedTerm = NVar String | NAbs String NamedTerm | NApp NamedTerm NamedTerm
deriving Show
data Term = Var Integer | Abs Term | App Term Term
deriving Show
class Scope s a where
find :: a -> s -> Maybe Integer
push :: a -> s -> s
pop :: s -> (a, s)
findAt :: Integer -> s -> a
newtype NamingContext a = NamingContext [a]
deriving Show
instance (Eq a) => Scope (NamingContext a) a where
find x (NamingContext ctx) = fmap toInteger $ elemIndex x ctx
push x (NamingContext ctx) = NamingContext (x : ctx)
pop (NamingContext ctx) = (head ctx, NamingContext $ tail ctx)
findAt x (NamingContext ctx) = genericIndex ctx x
type Context = NamingContext String
freeVariables :: NamedTerm -> NamingContext String
freeVariables = NamingContext . toList . freeVars
where freeVars :: NamedTerm -> Set String
freeVars (NVar x) = singleton x
freeVars (NAbs x t) = Data.Set.delete x $ freeVars t
freeVars (NApp s t) = Data.Set.union (freeVars s) (freeVars t)
unnameTerm :: NamedTerm -> (Context, Term)
unnameTerm nt = (ctx, term)
where ctx = freeVariables nt
term = uTerm ctx nt
uTerm :: Context -> NamedTerm -> Term
uTerm c (NVar x) = case find x c of
Just index -> Var index
Nothing -> error $ "Could not find variable " ++ x ++ " in scope"
uTerm c (NAbs x t) = Abs $ uTerm (push x c) t
uTerm c (NApp s t) = App (uTerm c s) (uTerm c t)
defaultContext :: Context
defaultContext = NamingContext vars
where vars = let numbers = [1..] :: [Integer]
snumbers = "" : map show numbers
letters = enumFromTo 'a' 'z' in
do
number <- snumbers
letter <- letters
return $ [letter] ++ number
diff :: Context -> Context -> Context
diff (NamingContext xs) (NamingContext ys) = NamingContext $ ys \\ xs
nameTerm :: Context -> Term -> NamedTerm
nameTerm ctx t = nameTerm' ctx (diff ctx defaultContext) t
where nameTerm' :: Context -> Context -> Term -> NamedTerm
nameTerm' c _ (Var x) = NVar (findAt x c)
nameTerm' c nextVar (Abs s) = NAbs x (nameTerm' ctx' nextVar' s)
where (x, nextVar') = pop nextVar
ctx' = push x c
nameTerm' c nextVar (App s u) = NApp (nameTerm' c nextVar s) (nameTerm' c nextVar u)
evalNamedTerm :: NamedTerm -> NamedTerm
evalNamedTerm = (uncurry nameTerm) . (second eval) . unnameTerm
eval :: Term -> Term
eval (Var x) = Var x
eval (Abs t) = Abs t
eval (App (Var s) t) = App (Var s) $ eval t
eval (App (Abs t) s) = eval $ shift (-1) $ substitute 0 (shift 1 (eval s)) t
eval (App (App u v) s) = case eval $ App u v of
App x y -> App (App x y) $ eval s
t -> eval $ App t s
substitute :: Integer -> Term -> Term -> Term
substitute k v (Var x) = if k == x then v else Var x
substitute k v (Abs t) = Abs $ substitute (k + 1) (shift 1 v) t
substitute k v (App a b) = App (substitute k v a) (substitute k v b)
shift :: Integer -> Term -> Term
shift m t = shift' m 0 t
where shift' :: Integer -> Integer -> Term -> Term
shift' n c (Var x) = if x < c then Var x else Var $ x + n
shift' n c (App s u) = App (shift' n c s) (shift' n c u)
shift' n c (Abs u) = Abs (shift' n (succ c) u)
| robertclancy/tapl | lambda/src/Lambda.hs | gpl-2.0 | 4,191 | 0 | 13 | 1,638 | 1,527 | 778 | 749 | 78 | 4 |
{-# LANGUAGE TemplateHaskell, DeriveDataTypeable, MultiParamTypeClasses #-}
module String_Matching.BM.Instance where
import String_Matching.Option
import Autolib.TES.Identifier
import Autolib.ToDoc
import Autolib.Reader
import Autolib.Set
import Data.Typeable
data Ord a => Instance a =
Instance { alphabet :: [ a ]
-- ^ as List because we need ordering for lambda
, word :: [ Option a ]
, bad_character :: [ Option Int]
, good_suffix :: [ Option Int ]
-- ^ this is m - \gamma (as in CLR)
}
deriving ( Eq, Typeable )
$(derives [makeReader, makeToDoc] [''Instance])
instance Ord a => Sub ( Instance a ) ( Instance a ) where
sub i j = and
[ sub ( word i ) ( word j )
, sub ( bad_character i ) ( bad_character j )
, sub ( good_suffix i ) ( good_suffix j )
]
example :: Instance Identifier
example = Instance
{ alphabet = [ read "a", read "b", read "c" ]
, word = read "[ ? , ? , ? , ? , ? , b , a , ? ]"
, bad_character = read "[ ? , ? , 0 ]"
, good_suffix = read "[ ? , ? , 0 , ? , 1 , ? , ? , 0 ]"
}
-- local variables:
-- mode: haskell
-- end:
| Erdwolf/autotool-bonn | src/String_Matching/BM/Instance.hs | gpl-2.0 | 1,189 | 6 | 10 | 360 | 314 | 173 | 141 | 26 | 1 |
{- |
Module : $Header$
Description : Provers for propositional logic
Copyright : (c) Dominik Luecke, Uni Bremen 2007
License : GPLv2 or higher, see LICENSE.txt
Maintainer : [email protected]
Stability : experimental
Portability : portable
This is the connection of the SAT-Solver minisat to Hets
-}
module Propositional.Prove
( zchaffProver -- the zChaff II Prover
, propConsChecker
) where
import qualified Propositional.AS_BASIC_Propositional as AS_BASIC
import qualified Propositional.Conversions as Cons
import qualified Propositional.Conversions as PC
import qualified Propositional.Morphism as PMorphism
import qualified Propositional.ProverState as PState
import qualified Propositional.Sign as Sig
import Propositional.Sublogic (PropSL, top)
import Proofs.BatchProcessing
import qualified Logic.Prover as LP
import Interfaces.GenericATPState
import GUI.GenericATP
import Common.ProofTree
import Common.Utils
import qualified Common.AS_Annotation as AS_Anno
import qualified Common.OrderedMap as OMap
import qualified Common.Result as Result
import Control.Monad (when)
import qualified Control.Concurrent as Concurrent
import Data.List
import Data.Maybe
import Data.Time (TimeOfDay, timeToTimeOfDay, midnight)
import System.Directory
import System.Process
import System.Exit
-- * Prover implementation
zchaffHelpText :: String
zchaffHelpText = "Zchaff is a very fast SAT-Solver \n" ++
"No additional Options are available" ++
"for it!"
-- | the name of the prover
zchaffS :: String
zchaffS = "zchaff"
{- |
The Prover implementation.
Implemented are: a prover GUI, and both commandline prover interfaces.
-}
zchaffProver
:: LP.Prover Sig.Sign AS_BASIC.FORMULA PMorphism.Morphism PropSL ProofTree
zchaffProver = LP.mkAutomaticProver zchaffS top zchaffProveGUI
zchaffProveCMDLautomaticBatch
{- |
The Consistency Cheker.
-}
propConsChecker :: LP.ConsChecker Sig.Sign AS_BASIC.FORMULA PropSL
PMorphism.Morphism ProofTree
propConsChecker = LP.mkConsChecker zchaffS top consCheck
consCheck :: String -> LP.TacticScript
-> LP.TheoryMorphism Sig.Sign AS_BASIC.FORMULA PMorphism.Morphism ProofTree
-> [LP.FreeDefMorphism AS_BASIC.FORMULA PMorphism.Morphism]
-- ^ free definitions
-> IO (LP.CCStatus ProofTree)
consCheck thName _ tm _ =
case LP.tTarget tm of
LP.Theory sig nSens -> do
let axioms = getAxioms $ snd $ unzip $ OMap.toList nSens
thName_clean = basename thName ++ "_cc.zchaff.dimacs"
dimacsOutput <- PC.showDIMACSProblem (thName ++ "_cc")
sig axioms Nothing
tmpFile <- getTempFile dimacsOutput thName_clean
(exitCode, resultHf, _) <-
readProcessWithExitCode zchaffS [tmpFile] ""
return $ if exitCode /= ExitSuccess then LP.CCStatus
(ProofTree $ "error by call zchaff " ++ thName)
midnight Nothing
else LP.CCStatus (ProofTree resultHf) midnight
$ searchResult resultHf
where
getAxioms :: [LP.SenStatus AS_BASIC.FORMULA (LP.ProofStatus ProofTree)]
-> [AS_Anno.Named AS_BASIC.FORMULA]
getAxioms f = map (AS_Anno.makeNamed "consistency" . AS_Anno.sentence)
$ filter AS_Anno.isAxiom f
searchResult :: String -> Maybe Bool
searchResult hf = let ls = lines hf in
if any (isInfixOf reUNSAT) ls then Just False else
if any (isInfixOf reSAT) ls then Just True
else Nothing
-- ** GUI
{- |
Invokes the generic prover GUI.
-}
zchaffProveGUI :: String -- ^ theory name
-> LP.Theory Sig.Sign AS_BASIC.FORMULA ProofTree
-> [LP.FreeDefMorphism AS_BASIC.FORMULA PMorphism.Morphism]
-- ^ free definitions
-> IO [LP.ProofStatus ProofTree] -- ^ proof status for each goal
zchaffProveGUI thName th freedefs =
genericATPgui (atpFun thName) True (LP.proverName zchaffProver) thName th
freedefs emptyProofTree
{- |
Parses a given default tactic script into a
'Interfaces.GenericATPState.ATPTacticScript' if possible.
-}
parseZchaffTacticScript :: LP.TacticScript -> ATPTacticScript
parseZchaffTacticScript = parseTacticScript batchTimeLimit []
-- ** command line function
{- |
Implementation of 'Logic.Prover.proveCMDLautomaticBatch' which provides an
automatic command line interface to the zchaff prover.
zchaff specific functions are omitted by data type ATPFunctions.
-}
zchaffProveCMDLautomaticBatch ::
Bool -- ^ True means include proved theorems
-> Bool -- ^ True means save problem file
-> Concurrent.MVar (Result.Result [LP.ProofStatus ProofTree])
-- ^ used to store the result of the batch run
-> String -- ^ theory name
-> LP.TacticScript -- ^ default tactic script
-> LP.Theory Sig.Sign AS_BASIC.FORMULA ProofTree
-- ^ theory consisting of a signature and a list of Named sentences
-> [LP.FreeDefMorphism AS_BASIC.FORMULA PMorphism.Morphism]
-- ^ free definitions
-> IO (Concurrent.ThreadId, Concurrent.MVar ())
{- ^ fst: identifier of the batch thread for killing it
snd: MVar to wait for the end of the thread -}
zchaffProveCMDLautomaticBatch inclProvedThs saveProblem_batch resultMVar
thName defTS th freedefs =
genericCMDLautomaticBatch (atpFun thName) inclProvedThs saveProblem_batch
resultMVar (LP.proverName zchaffProver) thName
(parseZchaffTacticScript defTS) th freedefs emptyProofTree
{- |
Record for prover specific functions. This is used by both GUI and command
line interface.
-}
atpFun :: String -- Theory name
-> ATPFunctions Sig.Sign AS_BASIC.FORMULA PMorphism.Morphism ProofTree
PState.PropProverState
atpFun thName = ATPFunctions
{ initialProverState = PState.propProverState
, goalOutput = Cons.goalDIMACSProblem thName
, atpTransSenName = PState.transSenName
, atpInsertSentence = PState.insertSentence
, proverHelpText = zchaffHelpText
, runProver = runZchaff
, batchTimeEnv = "HETS_ZCHAFF_BATCH_TIME_LIMIT"
, fileExtensions = FileExtensions
{ problemOutput = ".dimacs"
, proverOutput = ".zchaff"
, theoryConfiguration = ".czchaff"}
, createProverOptions = createZchaffOptions }
{- |
Runs zchaff. zchaff is assumed to reside in PATH.
-}
runZchaff :: PState.PropProverState
{- logical part containing the input Sign and
axioms and possibly goals that have been proved
earlier as additional axioms -}
-> GenericConfig ProofTree
-- configuration to use
-> Bool
-- True means save DIMACS file
-> String
-- Name of the theory
-> AS_Anno.Named AS_BASIC.FORMULA
-- Goal to prove
-> IO (ATPRetval
, GenericConfig ProofTree
)
-- (retval, configuration with proof status and complete output)
runZchaff pState cfg saveDIMACS thName nGoal = do
prob <- Cons.goalDIMACSProblem thName pState nGoal []
let thName_clean =
basename thName ++ '_' : AS_Anno.senAttr nGoal ++ ".dimacs"
when saveDIMACS (writeFile thName_clean prob)
zFileName <- getTempFile prob thName_clean
(_, zchaffOut, _) <- readProcessWithExitCode zchaffS
(zFileName : createZchaffOptions cfg) ""
(res, usedAxs, tUsed) <- analyzeZchaff zchaffOut pState
let defaultProofStatus =
(LP.openProofStatus (AS_Anno.senAttr nGoal) zchaffS
emptyProofTree)
{LP.tacticScript = LP.TacticScript $ show
ATPTacticScript
{ tsTimeLimit = configTimeLimit cfg
, tsExtraOpts = []} }
(err, retval) = case res of
Right p -> (ATPSuccess,
defaultProofStatus
{LP.goalStatus = p
, LP.usedAxioms = filter
(/= AS_Anno.senAttr nGoal) usedAxs
, LP.proofTree = ProofTree zchaffOut })
Left a -> (a, defaultProofStatus)
catch (removeFile zFileName) (const $ return ())
return (err, cfg
{ proofStatus = retval
, resultOutput = [zchaffOut]
, timeUsed = tUsed })
-- | analysis of output
analyzeZchaff :: String -> PState.PropProverState
-> IO (Either ATPRetval LP.GoalStatus, [String], TimeOfDay)
analyzeZchaff str' pState =
let unsat = isInfixOf reUNSAT str'
sat = isInfixOf reSAT str'
timeLine = fromMaybe "0" $ stripPrefix reTIME str'
timeout = isInfixOf reEndto str' || isInfixOf reEndmo str'
time = calculateTime timeLine
usedAx = map AS_Anno.senAttr $ PState.initialAxioms pState
in return (case () of
_ | timeout -> Left ATPTLimitExceeded
_ | sat && not unsat -> Right LP.Disproved
_ | not sat && unsat -> Right $ LP.Proved True
_ -> Left $ ATPError "Internal error.", usedAx, time)
-- | Calculated the time need for the proof in seconds
calculateTime :: String -> TimeOfDay
calculateTime timeLine =
timeToTimeOfDay $ realToFrac (fromMaybe
(error $ "calculateTime " ++ timeLine) $ readMaybe timeLine
:: Double)
reUNSAT :: String
reUNSAT = "RESULT:\tUNSAT"
reSAT :: String
reSAT = "RESULT:\tSAT"
reTIME :: String
reTIME = "Total Run Time"
reEndto :: String
reEndto = "TIME OUT"
reEndmo :: String
reEndmo = "MEM OUT"
{- |
Creates a list of all options the zChaff prover runs with.
Only Option is the timelimit
-}
createZchaffOptions :: GenericConfig ProofTree -> [String]
createZchaffOptions cfg =
[show $ configTimeLimit cfg]
| nevrenato/Hets_Fork | Propositional/Prove.hs | gpl-2.0 | 10,207 | 6 | 22 | 2,861 | 1,861 | 1,000 | 861 | 175 | 4 |
{-# LANGUAGE GeneralizedNewtypeDeriving, TemplateHaskell #-}
{-# LANGUAGE FlexibleInstances, MultiParamTypeClasses #-}
module Lamdu.Sugar.Convert.Monad
( OuterScopeInfo(..), osiPos, osiScope
, RecursiveRef(..), rrDefI, rrDefType
, ScopeInfo(..), siRecordParams, siNullParams, siLetItems, siExtractPos, siFloatPos
, Context(..)
, scInferContext, scTopLevelExpr, scPostProcessRoot, siRecursiveRef, scConfig
, scScopeInfo, scDebugMonitors, scCacheFunctions
, scOutdatedDefinitions, scFrozenDeps
, cachedFunc
, ConvertM(..), run, convertOnce
, PositionInfo(..)
, convertSubexpression
, typeProtectedSetToVal, typeProtect, postProcessAssert, postProcessWith
) where
import qualified Control.Lens as Lens
import Control.Monad.Once (OnceT, MonadOnce(..), Typeable)
import Control.Monad.Trans.Reader (ReaderT, runReaderT)
import qualified Control.Monad.Trans.Reader as Reader
import Control.Monad.Transaction (MonadTransaction(..))
import Data.Property (Property)
import Hyper.Unify (UVar)
import qualified Lamdu.Cache as Cache
import Lamdu.Calc.Definition (Deps)
import Lamdu.Calc.Infer (InferState)
import qualified Lamdu.Calc.Term as V
import qualified Lamdu.Calc.Type as T
import qualified Lamdu.Data.Anchors as Anchors
import qualified Lamdu.Data.Ops as DataOps
import qualified Lamdu.Debug as Debug
import qualified Lamdu.Expr.IRef as ExprIRef
import Lamdu.Sugar.Config (Config)
import qualified Lamdu.Sugar.Convert.Input as Input
import qualified Lamdu.Sugar.Convert.PostProcess as PostProcess
import Lamdu.Sugar.Internal
import qualified Lamdu.Sugar.Types as Sugar
import Revision.Deltum.Transaction (Transaction)
import qualified Revision.Deltum.Transaction as Transaction
import Text.PrettyPrint.HughesPJClass (prettyShow)
import Lamdu.Prelude
type T = Transaction
data OuterScopeInfo m = OuterScopeInfo
{ _osiPos :: ExprIRef.HRef m # V.Term
, _osiScope :: V.Scope # UVar
}
Lens.makeLenses ''OuterScopeInfo
data RecursiveRef m = RecursiveRef
{ _rrDefI :: ExprIRef.DefI m
, _rrDefType :: Pure # T.Scheme
}
Lens.makeLenses ''RecursiveRef
data ScopeInfo m = ScopeInfo
{ _siRecordParams :: Map V.Var (Set [T.Tag])
, _siNullParams :: Set V.Var
, -- Each let item potentially has an inline action
_siLetItems :: Map V.Var (Sugar.EntityId -> Sugar.VarInline (T m))
-- TODO: siTagParamInfos needs a reverse-lookup map too
, _siExtractPos :: Maybe (OuterScopeInfo m)
, _siFloatPos :: Maybe (OuterScopeInfo m)
, -- The globals we artificially inject into the scope in order to
-- infer their type supporting mutual recursions
_siRecursiveRef :: Maybe (RecursiveRef m)
}
Lens.makeLenses ''ScopeInfo
newtype ConvertM m a = ConvertM (ReaderT (Context m) (OnceT (T m)) a)
deriving newtype (Functor, Applicative, Monad, MonadReader (Context m))
instance Monad m => MonadTransaction m (ConvertM m) where
transaction = ConvertM . lift . lift
data PositionInfo = BinderPos | ExpressionPos deriving Eq
data Context m = Context
{ _scInferContext :: InferState
, _scCodeAnchors :: Anchors.CodeAnchors m
, _scScopeInfo :: ScopeInfo m
, _scTopLevelExpr :: Ann (Input.Payload m) # V.Term
, -- Check whether the definition is valid after an edit,
-- so that can detach bad edits.
_scPostProcessRoot :: T m PostProcess.Result
, _scOutdatedDefinitions ::
Map V.Var (Sugar.DefinitionOutdatedType InternalName (T m) ())
, _scFrozenDeps :: Property (T m) Deps
, _scDebugMonitors :: Debug.Monitors
, _scCacheFunctions :: Cache.Functions
, _scConfig :: Config
, scConvertSubexpression ::
PositionInfo -> Ann (Input.Payload m) # V.Term ->
ConvertM m (ExpressionU EvalPrep m)
}
Lens.makeLenses ''Context
instance Anchors.HasCodeAnchors (Context m) m where codeAnchors = scCodeAnchors
cachedFunc :: Monad m => (Cache.Functions -> a) -> ConvertM m a
cachedFunc f = Lens.view scCacheFunctions <&> f
typeProtect :: Monad m => T m PostProcess.Result -> T m a -> T m (Maybe a)
typeProtect checkOk act =
do
((result, isOk), changes) <- (,) <$> act <*> checkOk & Transaction.fork
case isOk of
PostProcess.GoodExpr -> Just result <$ Transaction.merge changes
PostProcess.BadExpr _ -> pure Nothing
typeProtectedSetToVal ::
Monad m =>
ConvertM m
(ExprIRef.HRef m # V.Term -> ExprIRef.ValI m -> T m (ExprIRef.ValI m))
typeProtectedSetToVal =
Lens.view scPostProcessRoot
<&> \checkOk dest valI ->
do
mResult <- DataOps.replace dest valI & typeProtect checkOk
case mResult of
Just result -> pure result
Nothing ->
do
res <- DataOps.setToAppliedHole valI dest
_ <- checkOk
pure res
postProcessWith :: Monad m => ConvertM m ((Pure # T.TypeError -> T m ()) -> T m ())
postProcessWith =
Lens.view scPostProcessRoot
<&> \postProcess onError ->
postProcess
>>=
\case
PostProcess.GoodExpr -> pure ()
PostProcess.BadExpr err ->
do
onError err
postProcess >>=
\case
PostProcess.GoodExpr -> pure ()
PostProcess.BadExpr e -> error ("postProcessWith onError failed: " <> prettyShow e)
postProcessAssert :: Monad m => ConvertM m (T m ())
postProcessAssert = postProcessWith ?? error . prettyShow
run :: (HasCallStack, Monad m) => Context m -> ConvertM m a -> OnceT (T m) a
run ctx (ConvertM action) =
runReaderT action ctx & report
where
Debug.EvaluatorM report = ctx ^. scDebugMonitors . Debug.sugaring . Debug.mAction
convertOnce :: (Monad m, Typeable a) => ConvertM m a -> ConvertM m (OnceT (T m) a)
convertOnce action =
Lens.view id
<&> (`run` action)
>>= ConvertM . lift . once
convertSubexpression :: Monad m => Ann (Input.Payload m) # V.Term -> ConvertM m (ExpressionU EvalPrep m)
convertSubexpression exprI =
do
convertSub <- Lens.view id <&> scConvertSubexpression
convertSub ExpressionPos exprI
| lamdu/lamdu | src/Lamdu/Sugar/Convert/Monad.hs | gpl-3.0 | 6,248 | 0 | 20 | 1,440 | 1,747 | 957 | 790 | -1 | -1 |
module XMonad.Actions.Fizzixnerd where
import XMonad
import System.Desktop.Commands.Fizzixnerd
waitHalfSecond :: X ()
waitHalfSecond = do
spawn "sleep 0.5"
runBrowser :: X ()
runBrowser = do
spawn browser
runExplorer :: X ()
runExplorer = do
spawn explorer
runSystemMonitor :: X ()
runSystemMonitor = do
spawn systemMonitor
runVideo :: X ()
runVideo = do
spawn video
runMusic :: X ()
runMusic = do
spawn music
runTerminal :: X ()
runTerminal = do
spawn myTerminal
runEnableHorizontalTwoFingerScrolling :: X ()
runEnableHorizontalTwoFingerScrolling = do
spawn enableHorizontalTwoFingerScrolling
runCompton :: X ()
runCompton = do
spawn compton
runGnomeDo :: X ()
runGnomeDo = do
spawn gnomeDo
runDock :: X ()
runDock = do
spawn dock
runVolumeUp :: X ()
runVolumeUp = do
spawn volumeUp
runVolumeDown :: X ()
runVolumeDown = do
spawn volumeDown
runToggleMute :: X ()
runToggleMute = do
spawn toggleMute
| Fizzixnerd/xmonad-config | site-haskell/src/XMonad/Actions/Fizzixnerd.hs | gpl-3.0 | 944 | 0 | 7 | 178 | 311 | 152 | 159 | 45 | 1 |
{- Merch.Race.UI.DrawingCombinators - Wrapper for Graphics.DrawingCombinators
Copyright 2013 Alan Manuel K. Gloria
This file is part of Merchant's Race.
Merchant's Race is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
Merchant's Race is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with Merchant's Race. If not, see <http://www.gnu.org/licenses/>.
-}
{- Wrapper for Graphics.DrawingCombinators. -}
module Merch.Race.UI.DrawingCombinators
( Affine
, R
, R2
, compose
, apply
, identity
, translate
, rotate
, scale
, inverse
, (%%)
, Image
, render
, sample
, point
, line
, regularPoly
, circle
, convexPoly
, bezierCurve
, rectangle -- create a rectangle
, Color(..)
, modulate
, tint
-- Graphics.DrawingCombinators' sprites leave much
-- to be desired. In particular, sprites are assumed
-- to always have an aspect ratio of 1, or if not,
-- for the caller to know beforehand what the aspect
-- ratio is.
, Sprite
, openSprites
, openSprite
, spriteAspect
, sprite
, Font
, openFont
, text
, textWidth
-- Special operations
, invisible -- Don't render the image, but sample still works
, forceSample -- Force an image to have a particular sample on all points
, intersection -- Combine the samples of two (Image Any) using an
-- intersection instead of union (default in Monoid of Any).
, Monoid(..)
, Any(..)
) where
import qualified Codec.Image.STB as STB
import Control.Applicative
import Control.Monad
import qualified Data.Bitmap.Base as BMP
import Data.List
import Graphics.DrawingCombinators hiding (Sprite, sprite, openSprite)
import qualified Graphics.Rendering.OpenGL.GL as GL
import Graphics.Rendering.OpenGL.GL(($=))
import System.IO.Unsafe
invisible :: Image a -> Image a
invisible im = unsafeOpenGLImage (\_ -> return ()) (sample im)
forceSample :: a -> Image b -> Image a
forceSample a im = pure (const a) <*> im
intersection :: Image Any -> Image Any -> Image Any
intersection a b = pure anyAnd <*> a <*> b
where
anyAnd (Any a) (Any b) = Any $ a && b
{- Sprite reimplementation. -}
data Sprite
= Sprite
{ sTexture :: GL.TextureObject
-- texture coordinates.
, sLowerLeft :: (GL.GLdouble, GL.GLdouble)
, sUpperRight :: (GL.GLdouble, GL.GLdouble)
-- aspect ratio
, spriteAspect :: GL.GLdouble
}
openSprites :: [FilePath] -> IO [Sprite]
openSprites [] = return []
openSprites fs = do
let -- We will be sorting the sprite bitmaps
-- in order from largest to smallest,
-- so add an ordering id to ensure that
-- our outputs will be ordered the same
-- as the inputs.
ifs = zip [0..] fs :: [(Integer, FilePath)]
ibss <- forM ifs $ \ (i, f) -> do
eb <- STB.loadImage f
case eb of
Left e -> fail e
Right b -> return (i, b, BMP.bitmapSize b)
-- put the widest items first.
let sorted_ibss = sortBy (\ (_,_,a) (_,_,b) -> compare a b) ibss
(tex, itxlas) <- layout sorted_ibss
let iss :: [(Integer, Sprite)]
iss = map (\ (i, ll, ur, a) -> (i, Sprite tex ll ur a)) itxlas
sorted_iss = sortBy (\ (a, _) (b, _) -> compare a b) iss
return $ map snd sorted_iss
where
-- Degenerate case
layout [(i, b, s)] = do
let aspect = getAspect s
(w, h) = s
larger
| w > h = w
| otherwise = h
texDim = pow2 larger
undefined
layout (ibs:ibss) = do
undefined
-- Compute the aspect ratio
getAspect (w, h) = fromIntegral w / fromIntegral h :: GL.GLdouble
-- Find the next higher power of 2
pow2 = loop 1
where
loop l i
| l > i = l
| otherwise = loop (l * 2) i
openSprite :: FilePath -> IO Sprite
openSprite f = do
[s] <- openSprites [f]
return s
sprite :: Sprite -> Image Any
sprite s = unsafeOpenGLImage render pick
where
thisTexture = sTexture s
render _ = do
cur <- GL.get $ GL.textureBinding GL.Texture2D
change <- case cur of
Nothing -> return True
Just t -> return $ thisTexture /= t
when change $ do
GL.texture GL.Texture2D $= GL.Enabled
GL.textureBinding GL.Texture2D $= Just thisTexture
GL.renderPrimitive GL.Quads $ do
let (lox, loy) = sLowerLeft s
(hix, hiy) = sUpperRight s
texcoord lox loy
vertex (negate xr) (yr)
texcoord hix loy
vertex (xr) (yr)
texcoord hix hiy
vertex (xr) (negate yr)
texcoord lox hiy
vertex (negate xr) (negate yr)
texcoord x y = GL.texCoord $ GL.TexCoord2 x y
vertex x y = GL.vertex $ GL.Vertex2
(realToFrac x :: GL.GLdouble)
(realToFrac y :: GL.GLdouble)
aspect = realToFrac $ spriteAspect s
(xr, yr)
| aspect >= 1 = (aspect, 1 )
| otherwise = (1 , 1 / aspect)
pick (x,y)
| negate xr <= x && x <= xr && negate yr <= y && y <= yr = Any True
| otherwise = Any False
-------
rectangle :: (R, R) -> (R, R) -> Image Any
rectangle (x1, y1) (x2, y2)
= core (min x1 x2) (min y1 y2) (max x1 x2) (max y1 y2)
where
core lx ly ux uy
= convexPoly
[ (lx, ly)
, (ux, ly)
, (ux, uy)
, (lx, uy)
]
| AmkG/merchants-race | Merch/Race/UI/DrawingCombinators.hs | gpl-3.0 | 5,655 | 0 | 18 | 1,534 | 1,646 | 882 | 764 | 138 | 3 |
module AutocorrectorSpec (spec) where
import Test.Hspec
import Language.Mulang
import Language.Mulang.Analyzer.Autocorrector
import Language.Mulang.Analyzer.Analysis hiding (spec)
import qualified Language.Mulang.Analyzer.Analysis as A
import Data.Maybe (fromJust)
import qualified Data.Map as Map
transform = head . fromJust . expectations . A.spec . autocorrect
run language = transform . expectationsAnalysis (CodeSample language "foo") . (:[])
runWithCustomRules language code rules e = transform (Analysis
code
(emptyAnalysisSpec { originalLanguage = language, expectations = Just [e], autocorrectionRules = Just $ Map.fromList rules }))
spec :: Spec
spec = do
describe "correct custom usages" $ do
it "corrects given rules when using MulangSample and originalLanguage" $ do
let setting = runWithCustomRules (Just Ruby) (MulangSample (Just None)) [("Uses:foo", "UsesPlus")]
setting (Expectation "*" "UsesPlus") `shouldBe` (Expectation "*" "UsesPlus")
setting (Expectation "*" "Uses:foo") `shouldBe` (Expectation "*" "UsesPlus")
it "corrects given rules when using MulangSample and no originalLanguage" $ do
let setting = runWithCustomRules Nothing (MulangSample (Just None)) [("Uses:foo", "UsesPlus")]
setting (Expectation "*" "UsesPlus") `shouldBe` (Expectation "*" "UsesPlus")
setting (Expectation "*" "Uses:foo") `shouldBe` (Expectation "*" "UsesPlus")
it "corrects given rules when using CodeSample" $ do
let setting = runWithCustomRules Nothing (CodeSample JavaScript "x") [("Uses:foo", "UsesPlus")]
setting (Expectation "*" "UsesPlus") `shouldBe` (Expectation "*" "UsesPlus")
setting (Expectation "*" "Uses:foo") `shouldBe` (Expectation "*" "UsesPlus")
describe "correct primitive usages" $ do
it "never corrects plain *" $ do
let uses = Expectation "*" "Uses:*"
run Mulang uses `shouldBe` uses
run Ruby uses `shouldBe` uses
run JavaScript uses `shouldBe` uses
run Prolog uses `shouldBe` uses
run Php uses `shouldBe` uses
it "always corrects plain * with strict match " $ do
let strictUses = Expectation "*" "Uses:=*"
let usesMultiply = Expectation "*" "UsesMultiply"
run Mulang strictUses `shouldBe` strictUses
run Ruby strictUses `shouldBe` usesMultiply
run JavaScript strictUses `shouldBe` usesMultiply
run Prolog strictUses `shouldBe` usesMultiply
run Php strictUses `shouldBe` usesMultiply
it "corrects operators when strict match" $ do
run Java (Expectation "*" "Uses:=+") `shouldBe` (Expectation "*" "UsesPlus")
run Haskell (Expectation "*" "Uses:=-") `shouldBe` (Expectation "*" "UsesMinus")
run JavaScript (Expectation "*" "Uses:====") `shouldBe` (Expectation "*" "UsesEqual")
run JavaScript (Expectation "*" "Uses:===") `shouldBe` (Expectation "*" "UsesSimilar")
it "corrects haskell otherwise negated" $ do
run Haskell (Expectation "*" "Not:Uses:otherwise") `shouldBe` (Expectation "*" "Not:UsesOtherwise")
it "corrects haskell otherwise" $ do
run Haskell (Expectation "*" "Uses:otherwise") `shouldBe` (Expectation "*" "UsesOtherwise")
it "corrects haskell and" $ do
run Haskell (Expectation "*" "Uses:and") `shouldBe` (Expectation "*" "Uses:and")
run Haskell (Expectation "*" "Uses:&&") `shouldBe` (Expectation "*" "UsesAnd")
it "corrects haskell or" $ do
run Haskell (Expectation "*" "Uses:or") `shouldBe` (Expectation "*" "Uses:or")
run Haskell (Expectation "*" "Uses:||")`shouldBe` (Expectation "*" "UsesOr")
it "corrects haskell not" $ do
run Haskell (Expectation "*" "Uses:not") `shouldBe` (Expectation "*" "UsesNegation")
run Haskell (Expectation "*" "Uses:!") `shouldBe` (Expectation "*" "Uses:!")
it "corrects java and" $ do
run Java (Expectation "*" "Uses:and") `shouldBe` (Expectation "*" "Uses:and")
run Java (Expectation "*" "Uses:&&") `shouldBe` (Expectation "*" "UsesAnd")
run Java (Expectation "*" "Uses:%") `shouldBe` (Expectation "*" "UsesModulo")
run Java (Expectation "*" "Uses:&") `shouldBe` (Expectation "*" "UsesBitwiseAnd")
run Java (Expectation "*" "Uses:|") `shouldBe` (Expectation "*" "UsesBitwiseOr")
run Java (Expectation "*" "Uses:>>") `shouldBe` (Expectation "*" "UsesBitwiseRightShift")
run Java (Expectation "*" "Uses:<<") `shouldBe` (Expectation "*" "UsesBitwiseLeftShift")
it "corrects python and" $ do
run Python (Expectation "*" "Uses:and") `shouldBe` (Expectation "*" "UsesAnd")
run Python (Expectation "*" "Uses:&&") `shouldBe` (Expectation "*" "Uses:&&")
it "corrects ruby and" $ do
run Ruby (Expectation "*" "Uses:and") `shouldBe` (Expectation "*" "UsesAnd")
run Ruby (Expectation "*" "Uses:&&") `shouldBe` (Expectation "*" "UsesAnd")
it "corrects ruby size" $ do
run Ruby (Expectation "*" "Uses:size") `shouldBe` (Expectation "*" "UsesSize")
run Ruby (Expectation "*" "Uses:length") `shouldBe` (Expectation "*" "UsesSize")
it "corrects JS operators" $ do
run JavaScript (Expectation "*" "Uses:+") `shouldBe` (Expectation "*" "UsesPlus")
run JavaScript (Expectation "*" "Uses:&&") `shouldBe` (Expectation "*" "UsesAnd")
run JavaScript (Expectation "*" "Uses:||") `shouldBe` (Expectation "*" "UsesOr")
run JavaScript (Expectation "*" "Uses:%") `shouldBe` (Expectation "*" "UsesModulo")
run JavaScript (Expectation "*" "Uses:&") `shouldBe` (Expectation "*" "UsesBitwiseAnd")
run JavaScript (Expectation "*" "Uses:|") `shouldBe` (Expectation "*" "UsesBitwiseOr")
run JavaScript (Expectation "*" "Uses:>>") `shouldBe` (Expectation "*" "UsesBitwiseRightShift")
run JavaScript (Expectation "*" "Uses:<<") `shouldBe` (Expectation "*" "UsesBitwiseLeftShift")
run JavaScript (Expectation "*" "Uses:length") `shouldBe` (Expectation "*" "UsesSize")
it "corrects C operators" $ do
run C (Expectation "*" "Uses:+") `shouldBe` (Expectation "*" "UsesPlus")
run C (Expectation "*" "Uses:&&") `shouldBe` (Expectation "*" "UsesAnd")
run C (Expectation "*" "Uses:||") `shouldBe` (Expectation "*" "UsesOr")
run C (Expectation "*" "Uses:%") `shouldBe` (Expectation "*" "UsesModulo")
run C (Expectation "*" "Uses:&") `shouldBe` (Expectation "*" "UsesBitwiseAnd")
run C (Expectation "*" "Uses:|") `shouldBe` (Expectation "*" "UsesBitwiseOr")
run C (Expectation "*" "Uses:>>") `shouldBe` (Expectation "*" "UsesBitwiseRightShift")
run C (Expectation "*" "Uses:<<") `shouldBe` (Expectation "*" "UsesBitwiseLeftShift")
run C (Expectation "*" "Uses:^") `shouldBe` (Expectation "*" "UsesBitwiseXor")
describe "correct primitive declarations" $ do
it "corrects haskell otherwise negated" $ do
run Haskell (Expectation "*" "Not:Declares:otherwise") `shouldBe` (Expectation "*" "Not:DeclaresOtherwise")
it "corrects haskell otherwise" $ do
run Haskell (Expectation "*" "Declares:otherwise") `shouldBe` (Expectation "*" "DeclaresOtherwise")
it "corrects haskell and" $ do
run Haskell (Expectation "*" "Declares:and") `shouldBe` (Expectation "*" "Declares:and")
run Haskell (Expectation "*" "Declares:&&") `shouldBe` (Expectation "*" "DeclaresAnd")
it "corrects haskell or" $ do
run Haskell (Expectation "*" "Declares:or") `shouldBe` (Expectation "*" "Declares:or")
run Haskell (Expectation "*" "Declares:||")`shouldBe` (Expectation "*" "DeclaresOr")
it "corrects haskell not" $ do
run Haskell (Expectation "*" "Declares:not") `shouldBe` (Expectation "*" "DeclaresNegation")
run Haskell (Expectation "*" "Declares:!") `shouldBe` (Expectation "*" "Declares:!")
it "corrects java and" $ do
run Java (Expectation "*" "Declares:and") `shouldBe` (Expectation "*" "Declares:and")
run Java (Expectation "*" "Declares:&&") `shouldBe` (Expectation "*" "DeclaresAnd")
it "corrects python and" $ do
run Python (Expectation "*" "Declares:and") `shouldBe` (Expectation "*" "DeclaresAnd")
run Python (Expectation "*" "Declares:&&") `shouldBe` (Expectation "*" "Declares:&&")
it "corrects ruby and" $ do
run Ruby (Expectation "*" "Declares:and") `shouldBe` (Expectation "*" "DeclaresAnd")
run Ruby (Expectation "*" "Declares:&&") `shouldBe` (Expectation "*" "DeclaresAnd")
describe "corrects keyword usages" $ do
it "corrects haskell type usage with negation" $ do
run Haskell (Expectation "*" "Not:Uses:type") `shouldBe` (Expectation "*" "Not:DeclaresTypeAlias")
it "corrects haskell type usage" $ do
run Haskell (Expectation "*" "Uses:type") `shouldBe` (Expectation "*" "DeclaresTypeAlias")
it "corrects java if usage" $ do
run Java (Expectation "*" "Uses:if") `shouldBe` (Expectation "*" "UsesIf")
it "corrects java class usage" $ do
run Java (Expectation "*" "Uses:class") `shouldBe` (Expectation "*" "DeclaresClass")
it "corrects java interface usage" $ do
run Java (Expectation "*" "Uses:interface") `shouldBe` (Expectation "*" "DeclaresInterface")
it "corrects java for usage" $ do
run Java (Expectation "*" "Uses:for") `shouldBe` (Expectation "*" "UsesForLoop")
it "corrects python def usage" $ do
run Python (Expectation "*" "Uses:def") `shouldBe` (Expectation "*" "DeclaresComputation")
it "corrects ruby class usage" $ do
run Ruby (Expectation "*" "Uses:class") `shouldBe` (Expectation "*" "DeclaresClass")
it "corrects ruby include usage" $ do
run Ruby (Expectation "*" "Uses:include") `shouldBe` (Expectation "*" "Includes")
it "corrects ruby def usage" $ do
run Ruby (Expectation "*" "Uses:def") `shouldBe` (Expectation "*" "DeclaresComputation")
describe "corrects keyword declarations" $ do
it "corrects haskell type declaration with negation" $ do
run Haskell (Expectation "*" "Not:Declares:type") `shouldBe` (Expectation "*" "Not:DeclaresTypeAlias")
it "corrects haskell type declaration" $ do
run Haskell (Expectation "*" "Declares:type") `shouldBe` (Expectation "*" "DeclaresTypeAlias")
it "corrects java if declaration" $ do
run Java (Expectation "*" "Declares:if") `shouldBe` (Expectation "*" "UsesIf")
it "corrects java class declaration" $ do
run Java (Expectation "*" "Declares:class") `shouldBe` (Expectation "*" "DeclaresClass")
it "corrects java interface declaration" $ do
run Java (Expectation "*" "Declares:interface") `shouldBe` (Expectation "*" "DeclaresInterface")
it "corrects java for declaration" $ do
run Java (Expectation "*" "Declares:for") `shouldBe` (Expectation "*" "UsesForLoop")
it "corrects python def declaration" $ do
run Python (Expectation "*" "Declares:def") `shouldBe` (Expectation "*" "DeclaresComputation")
it "corrects ruby class declaration" $ do
run Ruby (Expectation "*" "Declares:class") `shouldBe` (Expectation "*" "DeclaresClass")
it "corrects ruby include declaration" $ do
run Ruby (Expectation "*" "Declares:include") `shouldBe` (Expectation "*" "Includes")
it "corrects ruby def declaration" $ do
run Ruby (Expectation "*" "Declares:def") `shouldBe` (Expectation "*" "DeclaresComputation")
| mumuki/mulang | spec/AutocorrectorSpec.hs | gpl-3.0 | 11,462 | 0 | 22 | 2,259 | 3,471 | 1,700 | 1,771 | 164 | 1 |
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE DeriveDataTypeable #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE NoImplicitPrelude #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TypeFamilies #-}
{-# LANGUAGE TypeOperators #-}
{-# OPTIONS_GHC -fno-warn-duplicate-exports #-}
{-# OPTIONS_GHC -fno-warn-unused-binds #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- |
-- Module : Network.Google.Resource.SQL.SSLCerts.List
-- Copyright : (c) 2015-2016 Brendan Hay
-- License : Mozilla Public License, v. 2.0.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : auto-generated
-- Portability : non-portable (GHC extensions)
--
-- Lists all of the current SSL certificates for the instance.
--
-- /See:/ <https://cloud.google.com/sql/docs/reference/latest Cloud SQL Administration API Reference> for @sql.sslCerts.list@.
module Network.Google.Resource.SQL.SSLCerts.List
(
-- * REST Resource
SSLCertsListResource
-- * Creating a Request
, sslCertsList
, SSLCertsList
-- * Request Lenses
, sclProject
, sclInstance
) where
import Network.Google.Prelude
import Network.Google.SQLAdmin.Types
-- | A resource alias for @sql.sslCerts.list@ method which the
-- 'SSLCertsList' request conforms to.
type SSLCertsListResource =
"sql" :>
"v1beta4" :>
"projects" :>
Capture "project" Text :>
"instances" :>
Capture "instance" Text :>
"sslCerts" :>
QueryParam "alt" AltJSON :>
Get '[JSON] SSLCertsListResponse
-- | Lists all of the current SSL certificates for the instance.
--
-- /See:/ 'sslCertsList' smart constructor.
data SSLCertsList = SSLCertsList'
{ _sclProject :: !Text
, _sclInstance :: !Text
} deriving (Eq,Show,Data,Typeable,Generic)
-- | Creates a value of 'SSLCertsList' with the minimum fields required to make a request.
--
-- Use one of the following lenses to modify other fields as desired:
--
-- * 'sclProject'
--
-- * 'sclInstance'
sslCertsList
:: Text -- ^ 'sclProject'
-> Text -- ^ 'sclInstance'
-> SSLCertsList
sslCertsList pSclProject_ pSclInstance_ =
SSLCertsList'
{ _sclProject = pSclProject_
, _sclInstance = pSclInstance_
}
-- | Project ID of the project for which to list Cloud SQL instances.
sclProject :: Lens' SSLCertsList Text
sclProject
= lens _sclProject (\ s a -> s{_sclProject = a})
-- | Cloud SQL instance ID. This does not include the project ID.
sclInstance :: Lens' SSLCertsList Text
sclInstance
= lens _sclInstance (\ s a -> s{_sclInstance = a})
instance GoogleRequest SSLCertsList where
type Rs SSLCertsList = SSLCertsListResponse
type Scopes SSLCertsList =
'["https://www.googleapis.com/auth/cloud-platform",
"https://www.googleapis.com/auth/sqlservice.admin"]
requestClient SSLCertsList'{..}
= go _sclProject _sclInstance (Just AltJSON)
sQLAdminService
where go
= buildClient (Proxy :: Proxy SSLCertsListResource)
mempty
| rueshyna/gogol | gogol-sqladmin/gen/Network/Google/Resource/SQL/SSLCerts/List.hs | mpl-2.0 | 3,248 | 0 | 15 | 786 | 388 | 233 | 155 | 64 | 1 |
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE DeriveDataTypeable #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE NoImplicitPrelude #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE TypeFamilies #-}
{-# LANGUAGE TypeOperators #-}
{-# OPTIONS_GHC -fno-warn-duplicate-exports #-}
{-# OPTIONS_GHC -fno-warn-unused-binds #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- |
-- Module : Network.Google.Resource.Compute.NodeTemplates.Delete
-- Copyright : (c) 2015-2016 Brendan Hay
-- License : Mozilla Public License, v. 2.0.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : auto-generated
-- Portability : non-portable (GHC extensions)
--
-- Deletes the specified NodeTemplate resource.
--
-- /See:/ <https://developers.google.com/compute/docs/reference/latest/ Compute Engine API Reference> for @compute.nodeTemplates.delete@.
module Network.Google.Resource.Compute.NodeTemplates.Delete
(
-- * REST Resource
NodeTemplatesDeleteResource
-- * Creating a Request
, nodeTemplatesDelete
, NodeTemplatesDelete
-- * Request Lenses
, ntdRequestId
, ntdProject
, ntdNodeTemplate
, ntdRegion
) where
import Network.Google.Compute.Types
import Network.Google.Prelude
-- | A resource alias for @compute.nodeTemplates.delete@ method which the
-- 'NodeTemplatesDelete' request conforms to.
type NodeTemplatesDeleteResource =
"compute" :>
"v1" :>
"projects" :>
Capture "project" Text :>
"regions" :>
Capture "region" Text :>
"nodeTemplates" :>
Capture "nodeTemplate" Text :>
QueryParam "requestId" Text :>
QueryParam "alt" AltJSON :> Delete '[JSON] Operation
-- | Deletes the specified NodeTemplate resource.
--
-- /See:/ 'nodeTemplatesDelete' smart constructor.
data NodeTemplatesDelete =
NodeTemplatesDelete'
{ _ntdRequestId :: !(Maybe Text)
, _ntdProject :: !Text
, _ntdNodeTemplate :: !Text
, _ntdRegion :: !Text
}
deriving (Eq, Show, Data, Typeable, Generic)
-- | Creates a value of 'NodeTemplatesDelete' with the minimum fields required to make a request.
--
-- Use one of the following lenses to modify other fields as desired:
--
-- * 'ntdRequestId'
--
-- * 'ntdProject'
--
-- * 'ntdNodeTemplate'
--
-- * 'ntdRegion'
nodeTemplatesDelete
:: Text -- ^ 'ntdProject'
-> Text -- ^ 'ntdNodeTemplate'
-> Text -- ^ 'ntdRegion'
-> NodeTemplatesDelete
nodeTemplatesDelete pNtdProject_ pNtdNodeTemplate_ pNtdRegion_ =
NodeTemplatesDelete'
{ _ntdRequestId = Nothing
, _ntdProject = pNtdProject_
, _ntdNodeTemplate = pNtdNodeTemplate_
, _ntdRegion = pNtdRegion_
}
-- | An optional request ID to identify requests. Specify a unique request ID
-- so that if you must retry your request, the server will know to ignore
-- the request if it has already been completed. For example, consider a
-- situation where you make an initial request and the request times out.
-- If you make the request again with the same request ID, the server can
-- check if original operation with the same request ID was received, and
-- if so, will ignore the second request. This prevents clients from
-- accidentally creating duplicate commitments. The request ID must be a
-- valid UUID with the exception that zero UUID is not supported
-- (00000000-0000-0000-0000-000000000000).
ntdRequestId :: Lens' NodeTemplatesDelete (Maybe Text)
ntdRequestId
= lens _ntdRequestId (\ s a -> s{_ntdRequestId = a})
-- | Project ID for this request.
ntdProject :: Lens' NodeTemplatesDelete Text
ntdProject
= lens _ntdProject (\ s a -> s{_ntdProject = a})
-- | Name of the NodeTemplate resource to delete.
ntdNodeTemplate :: Lens' NodeTemplatesDelete Text
ntdNodeTemplate
= lens _ntdNodeTemplate
(\ s a -> s{_ntdNodeTemplate = a})
-- | The name of the region for this request.
ntdRegion :: Lens' NodeTemplatesDelete Text
ntdRegion
= lens _ntdRegion (\ s a -> s{_ntdRegion = a})
instance GoogleRequest NodeTemplatesDelete where
type Rs NodeTemplatesDelete = Operation
type Scopes NodeTemplatesDelete =
'["https://www.googleapis.com/auth/cloud-platform",
"https://www.googleapis.com/auth/compute"]
requestClient NodeTemplatesDelete'{..}
= go _ntdProject _ntdRegion _ntdNodeTemplate
_ntdRequestId
(Just AltJSON)
computeService
where go
= buildClient
(Proxy :: Proxy NodeTemplatesDeleteResource)
mempty
| brendanhay/gogol | gogol-compute/gen/Network/Google/Resource/Compute/NodeTemplates/Delete.hs | mpl-2.0 | 4,735 | 0 | 17 | 1,072 | 552 | 330 | 222 | 87 | 1 |
{-# LANGUAGE DeriveDataTypeable #-}
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE NoImplicitPrelude #-}
{-# LANGUAGE OverloadedStrings #-}
{-# OPTIONS_GHC -fno-warn-unused-imports #-}
-- |
-- Module : Network.Google.Composer.Types
-- Copyright : (c) 2015-2016 Brendan Hay
-- License : Mozilla Public License, v. 2.0.
-- Maintainer : Brendan Hay <[email protected]>
-- Stability : auto-generated
-- Portability : non-portable (GHC extensions)
--
module Network.Google.Composer.Types
(
-- * Service Configuration
composerService
-- * OAuth Scopes
, cloudPlatformScope
-- * Status
, Status
, status
, sDetails
, sCode
, sMessage
-- * OperationSchema
, OperationSchema
, operationSchema
, osAddtional
-- * ListImageVersionsResponse
, ListImageVersionsResponse
, listImageVersionsResponse
, livrNextPageToken
, livrImageVersions
-- * ListEnvironmentsResponse
, ListEnvironmentsResponse
, listEnvironmentsResponse
, lerNextPageToken
, lerEnvironments
-- * WebServerConfig
, WebServerConfig
, webServerConfig
, wscMachineType
-- * DatabaseConfig
, DatabaseConfig
, databaseConfig
, dcMachineType
-- * EnvironmentConfig
, EnvironmentConfig
, environmentConfig
, ecDatabaseConfig
, ecWebServerConfig
, ecNodeConfig
, ecNodeCount
, ecPrivateEnvironmentConfig
, ecEncryptionConfig
, ecSoftwareConfig
, ecDagGcsPrefix
, ecWebServerNetworkAccessControl
, ecGkeCluster
, ecAirflowURI
-- * ListOperationsResponse
, ListOperationsResponse
, listOperationsResponse
, lorNextPageToken
, lorOperations
-- * NodeConfig
, NodeConfig
, nodeConfig
, ncDiskSizeGb
, ncLocation
, ncNetwork
, ncOAuthScopes
, ncIPAllocationPolicy
, ncServiceAccount
, ncSubnetwork
, ncMachineType
, ncTags
-- * Operation
, Operation
, operation
, oDone
, oError
, oResponse
, oName
, oMetadata
-- * Empty
, Empty
, empty
-- * OperationMetadataOperationType
, OperationMetadataOperationType (..)
-- * SoftwareConfigEnvVariables
, SoftwareConfigEnvVariables
, softwareConfigEnvVariables
, scevAddtional
-- * ImageVersion
, ImageVersion
, imageVersion
, ivUpgradeDisabled
, ivCreationDisabled
, ivReleaseDate
, ivImageVersionId
, ivSupportedPythonVersions
, ivIsDefault
-- * Environment
, Environment
, environment
, eState
, eConfig
, eUuid
, eUpdateTime
, eName
, eLabels
, eCreateTime
-- * SoftwareConfigAirflowConfigOverrides
, SoftwareConfigAirflowConfigOverrides
, softwareConfigAirflowConfigOverrides
, scacoAddtional
-- * StatusDetailsItem
, StatusDetailsItem
, statusDetailsItem
, sdiAddtional
-- * AllowedIPRange
, AllowedIPRange
, allowedIPRange
, airValue
, airDescription
-- * IPAllocationPolicy
, IPAllocationPolicy
, ipAllocationPolicy
, iapServicesSecondaryRangeName
, iapUseIPAliases
, iapClusterSecondaryRangeName
, iapClusterIPv4CIdRBlock
, iapServicesIPv4CIdRBlock
-- * CheckUpgradeResponseContainsPypiModulesConflict
, CheckUpgradeResponseContainsPypiModulesConflict (..)
-- * Date
, Date
, date
, dDay
, dYear
, dMonth
-- * CheckUpgradeResponsePypiDependencies
, CheckUpgradeResponsePypiDependencies
, checkUpgradeResponsePypiDependencies
, curpdAddtional
-- * SoftwareConfigPypiPackages
, SoftwareConfigPypiPackages
, softwareConfigPypiPackages
, scppAddtional
-- * OperationMetadataState
, OperationMetadataState (..)
-- * EnvironmentState
, EnvironmentState (..)
-- * Xgafv
, Xgafv (..)
-- * PrivateEnvironmentConfig
, PrivateEnvironmentConfig
, privateEnvironmentConfig
, pecWebServerIPv4CIdRBlock
, pecCloudSQLIPv4CIdRBlock
, pecWebServerIPv4ReservedRange
, pecPrivateClusterConfig
, pecEnablePrivateEnvironment
-- * SoftwareConfig
, SoftwareConfig
, softwareConfig
, scImageVersion
, scPythonVersion
, scPypiPackages
, scAirflowConfigOverrides
, scEnvVariables
-- * PrivateClusterConfig
, PrivateClusterConfig
, privateClusterConfig
, pccEnablePrivateEndpoint
, pccMasterIPv4CIdRBlock
, pccMasterIPv4ReservedRange
-- * EncryptionConfig
, EncryptionConfig
, encryptionConfig
, ecKmsKeyName
-- * CheckUpgradeResponse
, CheckUpgradeResponse
, checkUpgradeResponse
, curContainsPypiModulesConflict
, curBuildLogURI
, curImageVersion
, curPypiDependencies
, curPypiConflictBuildLogExtract
-- * WebServerNetworkAccessControl
, WebServerNetworkAccessControl
, webServerNetworkAccessControl
, wsnacAllowedIPRanges
-- * EnvironmentLabels
, EnvironmentLabels
, environmentLabels
, elAddtional
-- * OperationMetadata
, OperationMetadata
, operationMetadata
, omState
, omResourceUuid
, omResource
, omEndTime
, omOperationType
, omCreateTime
-- * OperationResponse
, OperationResponse
, operationResponse
, orAddtional
) where
import Network.Google.Composer.Types.Product
import Network.Google.Composer.Types.Sum
import Network.Google.Prelude
-- | Default request referring to version 'v1' of the Cloud Composer API. This contains the host and root path used as a starting point for constructing service requests.
composerService :: ServiceConfig
composerService
= defaultService (ServiceId "composer:v1")
"composer.googleapis.com"
-- | See, edit, configure, and delete your Google Cloud Platform data
cloudPlatformScope :: Proxy '["https://www.googleapis.com/auth/cloud-platform"]
cloudPlatformScope = Proxy
| brendanhay/gogol | gogol-composer/gen/Network/Google/Composer/Types.hs | mpl-2.0 | 5,988 | 0 | 7 | 1,444 | 624 | 432 | 192 | 177 | 1 |
import Data.List
-- needed for strict eval foldl, foldl'
------------------------------------------------------
------- 1. Find the last element of a list. ----------
------------------------------------------------------
last' :: [xs] -> xs
-- .74s - 2.40s - 6.05s
last' xs = head (reverse xs)
-- with pattern matching
-- .01s - .02s - .02s
last' [] = error "No end for empty lists"
last' [x] = x
last' (_:xs) = last' xs
-- with guards
-- .01s - .02s - .02s
last' (x:xs)
| null xs = x
| otherwise = last' xs
-- with case statements
-- .02s - .03s - .04s
last' xs = case xs of (x:[]) -> x
(x:xs) -> last' xs
-- other solutions from website
-- .01s - .02s - .04s
last' = foldr1 (flip const)
-- .72s - 1.39s - 5.29s
last' = head . reverse
-- .01s - .02s - .02s
last' = foldl1 (curry snd)
-- .7s - 1.4s - 7.96s
last' [] = error "No end for empty lists!"
last' x = x !! (length x - 1)
-- REAL HASKELL SOLUTION IS TO USE last FUNCTION
--
-- performance notes:
-- Both the pattern-matching and guard formulations
-- using tail-call optimization are the most performant
-- of my solutions
--
-- They are suprisingly more performant than !! indexing
-- with the length function. I think this is because
-- the length function is a killa, performance-wise
------------------------------------------------------
------------------------------------------------------
-----------------------------------------------------
---- 2. Find the last but one element of a list. ----
-----------------------------------------------------
almost_last :: [xs] -> xs
-- .01s - .02s - .03s
almost_last [] = error "empty list"
almost_last (x:[]) = error "just one item"
almost_last (x:xs:[]) = x
almost_last (_:xs) = almost_last xs
-- .68s - 1.38s - 4.48s
almost_last xs = head (reverse (take 2 (reverse xs)))
-- with low precedence right-associative function applicator
-- .74s - 1.35s - 4.94s
almost_last xs = head $ reverse $ take 2 $ reverse xs
-- takes way too long
almost_last (x:xs)
| null xs = error "just one item"
| length xs == 1 = x
| otherwise = almost_last xs
-- point free style
-- takes way too long
almost_last = head . reverse . take 2 . reverse
-- other solutions from website
-- .01s - .03s - .04
almost_last = last . init
-- .76s - 1.45s - 6.37s
almost_last x = reverse x !! 1
-- .01s - .02s - .02s
almost_last [x,_] = x
almost_last (_:xs) = almost_last xs
-- .01s - .02s - .02s
almost_last (x:(_:[])) = x
almost_last (_:xs) = almost_last xs
-- .68s - 1.4s - 6.59s
almost_last = head . tail . reverse
-- THE REAL HASKELL SOLUTION IS TO USE last . init
--
-- performance notes:
-- Again, the pattern-matching solutions with TCO
-- is among the most performant
--
-- Also, 'last . init' is performant
--
-- The solution that uses guards sucks but that's
-- probably because the 'length' function is called
-- for every call to almost_last'''
-----------------------------------------------------
-----------------------------------------------------
------------------------------------------------------------
----- 3. Find the K'th element of a list. (1 indexed) ------
------------------------------------------------------------
kthel :: [xs] -> Int -> xs
-- .01s - .01s - .01s
kthel [] _ = error "out of bounds"
kthel xs 1 = head xs
kthel (_:xs) n = kthel xs (n-1)
-- .01s - .01s - .01s
kthel xs n = xs !! (n - 1)
-- .02s - .03s - .04s
kthel xs n = last $ take n xs
-- .02s - .03s - .04s
-- MMmmmm, curry
kthel2 n = last . take n
kthel = flip kthel2
-- above equivalent to
kthel = flip (\n -> last . take n)
-- .01s - .02s - .03s
kthel2 n = head . drop (n - 1)
kthel = flip dafunc2
-- THE REAL HASKELL is to use the !! indexing operator
--
-- Performance notes:
-- TCO pattern-matching is most performant
-- but also as performant as '!!' indexing,
-- in this case
--
-- head . drop is better that last . take
------------------------------------------------------------
------------------------------------------------------------
-------------------------------------------------------
------- 4. Find the number of elements of a list. -----
-------------------------------------------------------
length' :: [xs] -> Integer
-- using list comprehensions
-- .02s - .03s - .04s
length' xs = sum [1 | _ <- xs]
-- using recursion
-- CAUSES STACK OVERFLOW
length' [] = 0
length' xs = 1 + length' (tail xs)
-- using explicit if else
-- CAUSES STACK OVERFLOW
length' xs = if null xs
then 0
else 1 + length' (tail xs)
-- using guards
-- CAUSES STACK OVERFLOW
length' xs
| null xs = 0
| otherwise = 1 + length' (tail xs)
-- using pattern matching
-- CAUSES STACK OVERFLOW
length' (x:[]) = 1
length' (x:xs) = 1 + length' xs
-- .03s - .04s - .06s
length' = fst . last . zip [1..]
-- from the internet
-- uses tail recursion
-- .01s - .02s - .03s
length' xs = len xs 0
where len [] l = l
len (x:xs) l = len xs (l+1)
-- from the internet
-- change all values to 1 and sum it
-- .03s - .05s - .06s
length' = sum . map (\_ -> 1)
-- .02s - .03s - .03s
length' xs = foldl step 0 xs where step acc x = acc + 1
-- equivalent to above but using lambda
-- .01s - .03s - .05sj
length' = foldl (\acc _ -> 1 + acc) 0
-- CAUSES STACK OVERFLOW
length' = foldr (\_ acc -> acc + 1) 0
-- strict version
-- .01s - .03s - .04s
length' = foldl' (\acc _ -> 1 + acc) 0
-- the real solution is to use the length function
--
-- performance notes:
-- Any explicitly recursive function that cannot be
-- TCO-ed causes a stack overflow
--
-- The explicit one that uses TCO is very performant,
-- however
--
-- Similarly, the tail-recursive foldl works but the
-- non-tail-recursive foldr does not work
-------------------------------------------------------
-------------------------------------------------------
-------------------------------------------------------
----------------- 5. Reverse a list ------------------
-------------------------------------------------------
reverse' :: (Num a) => [a] -> [a]
-- takes too long
reverse' [] = []
reverse' (x:[]) = [x]
reverse' xs = last xs : (reverse' (init xs))
-- .00s - .01s - .01s
reverse' = foldl' (\acc x -> x : acc) []
-- .00s - .01s - .01s
reverse' = foldl (flip (:)) []
-- takes too long
reverse' xs = foldr (\x acc -> acc ++ [x]) [] xs
-- .01s - .02s - .04s
reverse' xs = reverse'' xs []
where reverse'' [] ys = ys
reverse'' (x:xs) ys = reverse'' xs (x:ys)
-- the real solution is to use the reverse function
--
-- performance notes:
-- The foldl (which I think are tail-recursive)
-- are really the only function that will work
-- with large lists. It is about as performant
-- as 'reverse'
-- The manual TCO version is very performant as well
-------------------------------------------------------
-------------------------------------------------------
main = do
let ls = [1..5]
let lsl = [1, 2, 3, 4, 5, 6, 5, 1, 2, 3]
let wds = "haskell"
print (last lsl)
print (kthel lsl 8)
print (kthel' lsl 8)
| tonyfischetti/99-haskell-problems | Q1to10.hs | unlicense | 7,242 | 0 | 11 | 1,615 | 1,579 | 864 | 715 | 84 | 2 |
{-# LANGUAGE NoMonomorphismRestriction #-}
module DetermineTheType where
-- simple example
example = 1
_1a = (* 9) 6
_1b = head [(0,"doge"),(1,"kitteh")]
_1c = head [(0 :: Integer, "doge"),(1,"kitteh")]
_1d = if False then True else False
_1e = length [1,2,3,4,5]
_1f = (length [1,2,3,4]) > (length "TACOCAT")
_2 = w where
x = 5
y = x + 5
w = y * 10
_3 = z where
x = 5
y = x + 5
z y = y * 10
_4 = f where
x = 5
y = x + 5
f = 4 / y
_5 = f where
x = "Julie"
y = " <3 "
z = "Haskell"
f = x ++ y ++ z
| dmp1ce/Haskell-Programming-Exercises | Chapter 5/determineTheType.hs | unlicense | 530 | 0 | 8 | 155 | 275 | 162 | 113 | 26 | 2 |
module Problem021 where
import Data.List (group)
main =
print $ sum $ filter amicable [2..10000]
amicable x = x /= x' && sumd x' == x
where x' = sumd x
sumd x = product (map f1 gs) - product (map f2 gs)
where gs = map (\g -> (head g, length g)) $ group $ factors x
f1 (p, m) = (p ^ (m + 1) - 1) `div` (p - 1)
f2 (p, m) = p ^ m
factors x = f x primes []
where f x (p:ps) fs
| x == 1 = fs
| r == 0 = p:(factors q)
| otherwise = f x ps fs where (q, r) = x `quotRem` p
primes = 2 : 3 : filter (f primes) [5, 7 ..]
where f (p:ps) n = p * p > n || n `rem` p /= 0 && f ps n
| vasily-kartashov/playground | euler/problem-021.hs | apache-2.0 | 647 | 0 | 13 | 231 | 396 | 206 | 190 | 17 | 1 |
-- Copyright 2016 Google Inc. All Rights Reserved.
--
-- Licensed under the Apache License, Version 2.0 (the "License");
-- you may not use this file except in compliance with the License.
-- You may obtain a copy of the License at
--
-- http://www.apache.org/licenses/LICENSE-2.0
--
-- Unless required by applicable law or agreed to in writing, software
-- distributed under the License is distributed on an "AS IS" BASIS,
-- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-- See the License for the specific language governing permissions and
-- limitations under the License.
{-# LANGUAGE TemplateHaskell, GADTs, KindSignatures, TypeSynonymInstances, FlexibleInstances #-}
import Data.List
import qualified Data.Map.Strict as M
import Control.Monad
import Control.Monad.Operational
import Control.Monad.Loops
import Lens.Micro
import Lens.Micro.TH
import qualified Test.QuickCheck as Q
import qualified Test.QuickCheck.Gen as Q
import Debug.Trace
{- Actors Library -}
type ProcessId = Int
type Message m = (ProcessId, m)
data ActorInstruction s m :: * -> * where
GetState :: ActorInstruction s m s
SetState :: s -> ActorInstruction s m ()
SendMessage :: ProcessId -> m -> ActorInstruction s m ()
WaitForMessage :: ActorInstruction s m (Message m)
type ActorProgram s m a = Program (ActorInstruction s m) a
type Queues m = M.Map ProcessId [m]
data Actor s m =
Actor {
_actorProgram :: ActorProgram s m (),
_actorState :: s,
_actorQueues :: Queues m
}
makeLenses ''Actor
enqueue :: ProcessId -> m -> Queues m -> Queues m
enqueue pid m = M.insertWith (flip (++)) pid [m]
dequeue :: ProcessId -> Queues m -> Maybe (m, Queues m)
dequeue pid queues
| Just (m:ms) <- M.lookup pid queues = Just (m, update ms)
| otherwise = Nothing
where update [] = M.delete pid queues
update ms = M.insert pid ms queues
stepActor :: ProcessId -> Actor s m -> (Actor s m, Maybe (Message m))
stepActor sid (Actor program state queues) = stepActor' (view program)
where stepActor' (GetState :>>= is) = (Actor (is state) state queues, Nothing)
stepActor' (SetState x :>>= is) = (Actor (is ()) x queues, Nothing)
stepActor' (SendMessage pid m :>>= is) = (Actor (is ()) state queues, Just (pid, m))
stepActor' (WaitForMessage :>>= is)
| Just (m, queues') <- dequeue sid queues = (Actor (is (sid, m)) state queues', Nothing)
stepActor' _ = (Actor program state queues, Nothing)
step :: (ProcessId, ProcessId) -> [Actor s m] -> [Actor s m]
step (pid, sid) actors | Just (rid, m) <- message = actors' & ix rid . actorQueues %~ enqueue pid m
| otherwise = actors'
where (actor', message) = stepActor sid (actors !! pid)
actors' = actors & ix pid .~ actor'
run :: [(ProcessId, ProcessId)] -> [Actor s m] -> [[Actor s m]]
run pids actors = scanl (flip step) actors pids
simpleRun :: [(ProcessId, ProcessId)] -> [(ActorProgram s m (), s)] -> [[Actor s m]]
simpleRun pids pairs = run pids [Actor p s M.empty | (p, s) <- pairs]
{- Helper Functions for writing ActorPrograms -}
getState = singleton GetState
setState s = singleton (SetState s)
sendMessage pid m = singleton (SendMessage pid m)
waitForMessage = singleton WaitForMessage
modify f = do { s <- getState; setState (f s) }
get l = do { s <- getState; return (s ^. l) }
infix 4 .=, %=, +=, ++=
l .= x = modify (l .~ x)
l %= f = modify (l %~ f)
l += x = l %= (+x)
l ++= x = l %= (++x)
{- Pretty-Printing -}
instance (Show s, Show m) => Show (ActorInstruction s m a) where
show GetState = "_ <- getState"
show (SetState x) = "setState " ++ show x
show (SendMessage pid m) = "sendMessage " ++ show pid ++ " " ++ show m
show WaitForMessage = "_ <- waitForMessage"
instance (Show s, Show m, Show a) => Show (ProgramView (ActorInstruction s m) a) where
show (Return x) = "return " ++ show x
show (i :>>= is) = show i ++ "; ..."
class Debuggable a where
debug :: ProcessId -> a -> String
-- removes boring steps from the trace
cleanup :: [(ProcessId, ProcessId)] -> [[Actor s m]] -> ([(ProcessId, ProcessId)], [[Actor s m]])
cleanup pids steps = let (pids', steps') = unzip (filter keep (zip pids steps)) in (pids', steps' ++ [last steps])
where keep ((pid, sid), actors) | (GetState :>>= _) <- view ((actors !! pid) ^. actorProgram) = False
keep _ = True
pretty :: (Debuggable s, Show s, Show m) => [(ProcessId, ProcessId)] -> [[Actor s m]] -> String
pretty pids (t:ts) = unlines (prettyActors t : zipWith prettyStep (pids) ts)
where prettyStep pid actors = banner ("advance " ++ show pid) ++ "\n\n" ++ prettyActors actors
banner x = "=============\n" ++ x ++ "\n============="
prettyActors actors = unlines (zipWith prettyActor [0..] actors)
prettyActor pid (Actor program state queues) =
unlines [ "pid " ++ show pid ++ ":"
, " program: " ++ show (view program)
, " state: " ++ show state
, " debug: " ++ debug pid state
, " queues: " ++ prettyQueues (M.toList queues) ]
prettyQueues queues = "{" ++ intercalate ", " (map prettyQueue queues) ++ "}"
prettyQueue (pid, queue) = show pid ++ ": [" ++ intercalate ", " (map show queue) ++ "]"
prettyClean pids snaphots = uncurry pretty (cleanup pids snaphots)
{- Example -}
data Msg = RequestResource | ReleaseResource | Ack
deriving Show
data State = State {
_clock :: Int,
_lastSeenTimestamps :: M.Map ProcessId Int,
_requestQueue :: [(Int, ProcessId)] -- always sorted
}
deriving Show
makeLenses ''State
ownsTheResource myPid [] lastSeen = False
ownsTheResource myPid ((ts, pid):_) lastSeen =
pid == myPid && all (> ts) (M.elems lastSeen)
instance Debuggable State where
debug pid state = "owns = " ++ show (ownsTheResource pid rq lastSeen)
where rq = state ^. requestQueue
lastSeen = state ^. lastSeenTimestamps
addToRequestQueue ts pid = requestQueue %= insert (ts,pid)
rmFromRequestQueue pid = requestQueue %= filter isNotPid
where isNotPid (_, qid) = qid /= pid
waitForMessage' = do
clock += 1
(ts, m) <- waitForMessage
c <- get clock
when (ts > c) (clock .= ts)
return (ts, m)
iOwnTheResource myPid =
ownsTheResource myPid <$> get requestQueue <*> get lastSeenTimestamps
incrementClock = do
c <- get clock
clock .= c + 1
return (c + 1)
program :: ProcessId -> [ProcessId] -> ActorProgram State (Int, Msg) ()
program myPid otherPids = forever $ do
iOwn <- iOwnTheResource myPid
if iOwn
then do
rmFromRequestQueue myPid
ts <- incrementClock
forM_ otherPids $ \pid -> do
sendMessage pid (ts, ReleaseResource)
else do
ts <- incrementClock
addToRequestQueue ts myPid
forM_ otherPids $ \pid -> do
sendMessage pid (ts, RequestResource)
whileM_ (not <$> iOwnTheResource myPid) $ do
(pid, (ots, msg)) <- waitForMessage'
lastSeenTimestamps %= M.insert pid ots
case msg of
RequestResource -> do
addToRequestQueue ots pid
ts <- incrementClock
sendMessage pid (ts, Ack)
ReleaseResource -> do
rmFromRequestQueue pid
_ -> return ()
actors = [
(program 0 [1,2], State 0 (M.fromList [(1,-1), (2,-1)]) [(-2,0)]),
(program 1 [0,2], State 0 (M.fromList [(0,-1), (2,-1)]) [(-2,0)]),
(program 2 [0,1], State 0 (M.fromList [(0,-1), (1,-1)]) [(-2,0)])]
-- We define an alias for lists of pairs of pids that represent execution
-- sequences. (The first element of a pair is the id of the process to advance,
-- the second element is the process it should receive a message from if such a
-- message is waiting and the next execution step is to wait for a messa ge.) It
-- allows us to define custom instances for Arbitrary and Show.
newtype ExecutionSequence = ExecutionSequence [(ProcessId, ProcessId)]
instance Q.Arbitrary ExecutionSequence where
arbitrary = ExecutionSequence <$> Q.listOf ((,) <$> pids <*> pids)
where pids = Q.elements [0,1,2]
shrink (ExecutionSequence pids) = ExecutionSequence <$> Q.shrink pids
instance Show ExecutionSequence where
show (ExecutionSequence ns) = unlines [show ns, prettyClean ns (simpleRun ns actors)]
property1 (ExecutionSequence pids) = all property' (simpleRun pids actors)
where property' actors = length (filter id (zipWith owns [0..] actors)) <= 1
owns pid actor = ownsTheResource pid (actor ^. actorState.requestQueue) (actor ^. actorState.lastSeenTimestamps)
property2 (ExecutionSequence pids) = all property' (simpleRun pids actors)
where property' actors = not (all stuck actors)
stuck actor = M.null (actor ^. actorQueues) && isWaiting (view (actor ^. actorProgram))
isWaiting (WaitForMessage :>>= _) = True
isWaiting _ = False
allpids = 0:1:2:allpids
fairExecutionSequence = concatMap (\pid -> [(pid, 0), (pid, 1), (pid, 2)]) allpids
fairTrace = simpleRun fairExecutionSequence actors
owners = map extractOwner fairTrace
where extractOwner actors = fst <$> find isOwner (zip [0..] actors)
isOwner (pid, actor) = ownsTheResource pid (actor ^. actorState.requestQueue) (actor ^. actorState.lastSeenTimestamps)
property3 = and [Just n `elem` owners | n <- [0,1,2]]
main = do
Q.quickCheckWith args property1
Q.quickCheckWith args property2
print property3
where args = Q.stdArgs { Q.maxSuccess = 1000, Q.maxSize = 1000 }
| polux/snippets | Lamport.hs | apache-2.0 | 9,470 | 0 | 20 | 2,086 | 3,457 | 1,813 | 1,644 | -1 | -1 |
{-# LANGUAGE DeriveDataTypeable #-}
{-# LANGUAGE FlexibleContexts #-}
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE GADTs #-}
{-# LANGUAGE LambdaCase #-}
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE RecordWildCards #-}
{-# LANGUAGE RecursiveDo #-}
{-# LANGUAGE ScopedTypeVariables #-}
{-# LANGUAGE TupleSections #-}
{-# LANGUAGE TypeSynonymInstances #-}
{-# LANGUAGE UnicodeSyntax #-}
{-# LANGUAGE ViewPatterns #-}
import Control.Monad
import System.Log.Logger (Priority(DEBUG),rootLoggerName,setLevel,updateGlobalLogger)
import System.Random (randomRIO)
import Test.Framework
import Test.Framework.Providers.QuickCheck2
import Test.HUnit hiding (Path,Test)
import Test.QuickCheck.Instances ()
import Test.QuickCheck.Property (ioProperty)
import LogicGrowsOnTrees.Examples.MapColoring
main :: IO ()
main = do
updateGlobalLogger rootLoggerName (setLevel DEBUG)
defaultMain tests
tests :: [Test]
tests =
[testGroup "LogicGrowsOnTrees.Examples"
[testProperty name . ioProperty $ do
number_of_colors ← randomRIO (2,5)
number_of_countries ← randomRIO (3,7)
neighbor_probability ← randomRIO (0,1::Float)
neighbors ← fmap (concat . concat) $
forM [1..number_of_countries] $ \x →
forM [x+1..number_of_countries] $ \y → do
outcome ← randomRIO (0,1)
return $
if outcome > neighbor_probability
then [(x,y),(y,x)]
else []
let solutions =
computeSolutions
number_of_colors
number_of_countries
(\x y → (x,y) `elem` neighbors)
forM_ solutions $ \solution →
forM_ solution $ \(country_1,color_1) →
forM_ solution $ \(country_2,color_2) →
when ((country_1,country_2) `elem` neighbors) $
assertBool "neighbors have different colors" $ color_1 /= color_2
let correct_count = sum $ do
solution ← zip [1..] <$> replicateM (fromIntegral number_of_countries) [1..number_of_colors]
forM_ solution $ \(country_1,color_1) →
forM_ solution $ \(country_2,color_2) →
when ((country_1,country_2) `elem` neighbors) $
guard $ color_1 /= color_2
return 1
computeCount number_of_colors solutions @?= correct_count
return True
| (name,computeSolutions,computeCount) ←
[("coloringSolutions",coloringSolutions,curry (fromIntegral . length . snd))
,("coloringUniqueSolutions",coloringUniqueSolutions,
\number_of_colors →
sum
.
map (\solution →
let number_of_colors_used = maximum . fmap snd $ solution
in product [number_of_colors-number_of_colors_used+1..number_of_colors]
)
)
]
]
]
| gcross/LogicGrowsOnTrees | LogicGrowsOnTrees/tests/test-Examples.hs | bsd-2-clause | 3,236 | 0 | 27 | 1,112 | 714 | 391 | 323 | 69 | 2 |
{-# LANGUAGE TemplateHaskell, TypeSynonymInstances, FlexibleInstances,
OverloadedStrings #-}
{-# OPTIONS_GHC -fno-warn-orphans #-}
{-| Unittests for ganeti-htools.
-}
{-
Copyright (C) 2009, 2010, 2011, 2012, 2013 Google Inc.
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright notice,
this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in the
documentation and/or other materials provided with the distribution.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR
CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-}
module Test.Ganeti.Objects
( testObjects
, Node(..)
, genConfigDataWithNetworks
, genDisk
, genDiskWithChildren
, genEmptyCluster
, genInst
, genInstWithNets
, genValidNetwork
, genBitStringMaxLen
) where
import Test.QuickCheck
import qualified Test.HUnit as HUnit
import Control.Applicative
import Control.Monad
import Data.Char
import qualified Data.List as List
import qualified Data.Map as Map
import Data.Maybe (fromMaybe)
import qualified Data.Set as Set
import Data.Word (Word32)
import GHC.Exts (IsString(..))
import System.Time (ClockTime(..))
import qualified Text.JSON as J
import Test.Ganeti.Query.Language ()
import Test.Ganeti.SlotMap (genSlotLimit)
import Test.Ganeti.TestHelper
import Test.Ganeti.TestCommon
import Test.Ganeti.Types ()
import qualified Ganeti.Constants as C
import qualified Ganeti.ConstantUtils as CU
import Ganeti.Network
import Ganeti.Objects as Objects
import qualified Ganeti.Objects.BitArray as BA
import Ganeti.JSON
import Ganeti.Types
-- * Arbitrary instances
instance Arbitrary (Container DataCollectorConfig) where
arbitrary = do
let names = CU.toList C.dataCollectorNames
activations <- vector $ length names
timeouts <- vector $ length names
let configs = zipWith DataCollectorConfig activations timeouts
return GenericContainer {
fromContainer = Map.fromList $ zip names configs }
$(genArbitrary ''PartialNDParams)
instance Arbitrary Node where
arbitrary = Node <$> genFQDN <*> genFQDN <*> genFQDN
<*> arbitrary <*> arbitrary <*> arbitrary <*> genFQDN
<*> arbitrary <*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary <*> arbitrary <*> genFQDN <*> arbitrary
<*> (Set.fromList <$> genTags)
$(genArbitrary ''BlockDriver)
$(genArbitrary ''DiskMode)
instance Arbitrary LogicalVolume where
arbitrary = LogicalVolume <$> validName <*> validName
where
validName = -- we intentionally omit '.' and '-' to avoid forbidden names
listOf1 $ elements (['a'..'z'] ++ ['A'..'Z'] ++ ['0'..'9'] ++ "+_")
instance Arbitrary DiskLogicalId where
arbitrary = oneof [ LIDPlain <$> arbitrary
, LIDDrbd8 <$> genFQDN <*> genFQDN <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
, LIDFile <$> arbitrary <*> arbitrary
, LIDBlockDev <$> arbitrary <*> arbitrary
, LIDRados <$> arbitrary <*> arbitrary
]
-- | 'Disk' 'arbitrary' instance. Since we don't test disk hierarchy
-- properties, we only generate disks with no children (FIXME), as
-- generating recursive datastructures is a bit more work.
instance Arbitrary Disk where
arbitrary =
frequency [ (2, liftM RealDisk $ RealDiskData <$> arbitrary
<*> pure [] <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary)
, (1, liftM ForthcomingDisk $ ForthcomingDiskData <$> arbitrary
<*> pure [] <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary <*> arbitrary <*> arbitrary
<*> arbitrary)
]
-- FIXME: we should generate proper values, >=0, etc., but this is
-- hard for partial ones, where all must be wrapped in a 'Maybe'
$(genArbitrary ''PartialBeParams)
$(genArbitrary ''AdminState)
$(genArbitrary ''AdminStateSource)
$(genArbitrary ''PartialNicParams)
$(genArbitrary ''PartialNic)
instance Arbitrary ForthcomingInstanceData where
arbitrary =
ForthcomingInstanceData
-- name
<$> genMaybe genFQDN
-- primary node
<*> genMaybe genFQDN
-- OS
<*> genMaybe genFQDN
-- hypervisor
<*> arbitrary
-- hvparams
-- FIXME: add non-empty hvparams when they're a proper type
<*> pure (GenericContainer Map.empty)
-- beparams
<*> arbitrary
-- osparams
<*> pure (GenericContainer Map.empty)
-- osparams_private
<*> pure (GenericContainer Map.empty)
-- admin_state
<*> genMaybe arbitrary
-- admin_state_source
<*> genMaybe arbitrary
-- nics
<*> arbitrary
-- disks
<*> vectorOf 5 arbitrary
-- disks active
<*> genMaybe arbitrary
-- network port
<*> arbitrary
-- ts
<*> arbitrary <*> arbitrary
-- uuid
<*> arbitrary
-- serial
<*> arbitrary
-- tags
<*> (Set.fromList <$> genTags)
instance Arbitrary RealInstanceData where
arbitrary =
RealInstanceData
-- name
<$> genFQDN
-- primary node
<*> genFQDN
-- OS
<*> genFQDN
-- hypervisor
<*> arbitrary
-- hvparams
-- FIXME: add non-empty hvparams when they're a proper type
<*> pure (GenericContainer Map.empty)
-- beparams
<*> arbitrary
-- osparams
<*> pure (GenericContainer Map.empty)
-- osparams_private
<*> pure (GenericContainer Map.empty)
-- admin_state
<*> arbitrary
-- admin_state_source
<*> arbitrary
-- nics
<*> arbitrary
-- disks
<*> vectorOf 5 arbitrary
-- disks active
<*> arbitrary
-- network port
<*> arbitrary
-- ts
<*> arbitrary <*> arbitrary
-- uuid
<*> arbitrary
-- serial
<*> arbitrary
-- tags
<*> (Set.fromList <$> genTags)
instance Arbitrary Instance where
arbitrary = frequency [ (1, ForthcomingInstance <$> arbitrary)
, (3, RealInstance <$> arbitrary)
]
-- | Generates an instance that is connected to the given networks
-- and possibly some other networks
genInstWithNets :: [String] -> Gen Instance
genInstWithNets nets = do
plain_inst <- RealInstance <$> arbitrary
enhanceInstWithNets plain_inst nets
-- | Generates an instance that is connected to some networks
genInst :: Gen Instance
genInst = genInstWithNets []
-- | Enhances a given instance with network information, by connecting it to the
-- given networks and possibly some other networks
enhanceInstWithNets :: Instance -> [String] -> Gen Instance
enhanceInstWithNets inst nets = do
mac <- arbitrary
ip <- arbitrary
nicparams <- arbitrary
name <- arbitrary
uuid <- arbitrary
-- generate some more networks than the given ones
num_more_nets <- choose (0,3)
more_nets <- vectorOf num_more_nets genUUID
let genNic net = PartialNic mac ip nicparams net name uuid
partial_nics = map (genNic . Just)
(List.nub (nets ++ more_nets))
new_inst = case inst of
RealInstance rinst ->
RealInstance rinst { realInstNics = partial_nics }
ForthcomingInstance _ -> inst
return new_inst
genDiskWithChildren :: Int -> Gen Disk
genDiskWithChildren num_children = do
logicalid <- arbitrary
children <- vectorOf num_children (genDiskWithChildren 0)
nodes <- arbitrary
ivname <- genName
size <- arbitrary
mode <- arbitrary
name <- genMaybe genName
spindles <- arbitrary
params <- arbitrary
uuid <- genName
serial <- arbitrary
time <- arbitrary
return . RealDisk $
RealDiskData logicalid children nodes ivname size mode name
spindles params uuid serial time time
genDisk :: Gen Disk
genDisk = genDiskWithChildren 3
-- | FIXME: This generates completely random data, without normal
-- validation rules.
$(genArbitrary ''PartialISpecParams)
-- | FIXME: This generates completely random data, without normal
-- validation rules.
$(genArbitrary ''PartialIPolicy)
$(genArbitrary ''FilledISpecParams)
$(genArbitrary ''MinMaxISpecs)
$(genArbitrary ''FilledIPolicy)
$(genArbitrary ''IpFamily)
$(genArbitrary ''FilledNDParams)
$(genArbitrary ''FilledNicParams)
$(genArbitrary ''FilledBeParams)
-- | No real arbitrary instance for 'ClusterHvParams' yet.
instance Arbitrary ClusterHvParams where
arbitrary = return $ GenericContainer Map.empty
-- | No real arbitrary instance for 'OsHvParams' yet.
instance Arbitrary OsHvParams where
arbitrary = return $ GenericContainer Map.empty
-- | No real arbitrary instance for 'GroupDiskParams' yet.
instance Arbitrary GroupDiskParams where
arbitrary = return $ GenericContainer Map.empty
instance Arbitrary ClusterNicParams where
arbitrary = (GenericContainer . Map.singleton C.ppDefault) <$> arbitrary
instance Arbitrary OsParams where
arbitrary = (GenericContainer . Map.fromList) <$> arbitrary
instance Arbitrary Objects.ClusterOsParamsPrivate where
arbitrary = (GenericContainer . Map.fromList) <$> arbitrary
instance Arbitrary a => Arbitrary (Private a) where
arbitrary = Private <$> arbitrary
instance Arbitrary ClusterOsParams where
arbitrary = (GenericContainer . Map.fromList) <$> arbitrary
instance Arbitrary ClusterBeParams where
arbitrary = (GenericContainer . Map.fromList) <$> arbitrary
instance Arbitrary TagSet where
arbitrary = Set.fromList <$> genTags
instance Arbitrary IAllocatorParams where
arbitrary = return $ GenericContainer Map.empty
$(genArbitrary ''Cluster)
instance Arbitrary ConfigData where
arbitrary = genEmptyCluster 0 >>= genConfigDataWithNetworks
instance Arbitrary AddressPool where
arbitrary = AddressPool . BA.fromList <$> arbitrary
instance Arbitrary Network where
arbitrary = genValidNetwork
instance Arbitrary FilterAction where
arbitrary = oneof
[ pure Accept
, pure Pause
, pure Reject
, pure Continue
, RateLimit <$> genSlotLimit
]
instance Arbitrary FilterPredicate where
arbitrary = oneof
[ FPJobId <$> arbitrary
, FPOpCode <$> arbitrary
, FPReason <$> arbitrary
]
instance Arbitrary FilterRule where
arbitrary = FilterRule <$> arbitrary
<*> arbitrary
<*> arbitrary
<*> arbitrary
<*> arbitrary
<*> genUUID
instance Arbitrary MaintenanceData where
arbitrary = MaintenanceData <$> (fromPositive <$> arbitrary)
<*> arbitrary
<*> arbitrary
<*> arbitrary
<*> arbitrary
-- | Generates a network instance with minimum netmasks of /24. Generating
-- bigger networks slows down the tests, because long bit strings are generated
-- for the reservations.
genValidNetwork :: Gen Objects.Network
genValidNetwork = do
-- generate netmask for the IPv4 network
netmask <- fromIntegral <$> choose (24::Int, 30)
name <- genName >>= mkNonEmpty
mac_prefix <- genMaybe genName
net <- arbitrary
net6 <- genMaybe genIp6Net
gateway <- genMaybe arbitrary
gateway6 <- genMaybe genIp6Addr
res <- liftM Just (genBitString $ netmask2NumHosts netmask)
ext_res <- liftM Just (genBitString $ netmask2NumHosts netmask)
uuid <- arbitrary
ctime <- arbitrary
mtime <- arbitrary
let n = Network name mac_prefix (mkIp4Network net netmask) net6 gateway
gateway6 res ext_res uuid ctime mtime 0 Set.empty
return n
-- | Generate an arbitrary string consisting of '0' and '1' of the given length.
genBitString :: Int -> Gen AddressPool
genBitString len =
(AddressPool . BA.fromList) `liftM` vectorOf len (elements [False, True])
-- | Generate an arbitrary string consisting of '0' and '1' of the maximum given
-- length.
genBitStringMaxLen :: Int -> Gen AddressPool
genBitStringMaxLen maxLen = choose (0, maxLen) >>= genBitString
-- | Generator for config data with an empty cluster (no instances),
-- with N defined nodes.
genEmptyCluster :: Int -> Gen ConfigData
genEmptyCluster ncount = do
nodes <- vector ncount
version <- arbitrary
grp <- arbitrary
let guuid = groupUuid grp
nodes' = zipWith (\n idx ->
let newname = takeWhile (/= '.') (nodeName n)
++ "-" ++ show idx
in (newname, n { nodeGroup = guuid,
nodeName = newname}))
nodes [(1::Int)..]
nodemap = Map.fromList nodes'
contnodes = if Map.size nodemap /= ncount
then error ("Inconsistent node map, duplicates in" ++
" node name list? Names: " ++
show (map fst nodes'))
else GenericContainer nodemap
continsts = GenericContainer Map.empty
networks = GenericContainer Map.empty
disks = GenericContainer Map.empty
filters = GenericContainer Map.empty
maintenance <- arbitrary
let contgroups = GenericContainer $ Map.singleton guuid grp
serial <- arbitrary
-- timestamp fields
ctime <- arbitrary
mtime <- arbitrary
cluster <- resize 8 arbitrary
let c = ConfigData version cluster contnodes contgroups continsts networks
disks filters ctime maintenance mtime serial
return c
-- | FIXME: make an even simpler base version of creating a cluster.
-- | Generates config data with a couple of networks.
genConfigDataWithNetworks :: ConfigData -> Gen ConfigData
genConfigDataWithNetworks old_cfg = do
num_nets <- choose (0, 3)
-- generate a list of network names (no duplicates)
net_names <- genUniquesList num_nets genName >>= mapM mkNonEmpty
-- generate a random list of networks (possibly with duplicate names)
nets <- vectorOf num_nets genValidNetwork
-- use unique names for the networks
let nets_unique = map ( \(name, net) -> net { networkName = name } )
(zip net_names nets)
net_map = GenericContainer $ Map.fromList
(map (\n -> (networkUuid n, n)) nets_unique)
new_cfg = old_cfg { configNetworks = net_map }
return new_cfg
-- * Test properties
-- | Tests that fillDict behaves correctly
prop_fillDict :: [(Int, Int)] -> [(Int, Int)] -> Property
prop_fillDict defaults custom =
let d_map = Map.fromList defaults
d_keys = map fst defaults
c_map = Map.fromList custom
c_keys = map fst custom
in conjoin [ counterexample "Empty custom filling"
(fillDict d_map Map.empty [] == d_map)
, counterexample "Empty defaults filling"
(fillDict Map.empty c_map [] == c_map)
, counterexample "Delete all keys"
(fillDict d_map c_map (d_keys++c_keys) == Map.empty)
]
prop_LogicalVolume_serialisation :: LogicalVolume -> Property
prop_LogicalVolume_serialisation = testSerialisation
prop_LogicalVolume_deserialisationFail :: Property
prop_LogicalVolume_deserialisationFail =
conjoin . map (testDeserialisationFail (LogicalVolume "" "")) $
[ J.JSArray []
, J.JSString $ J.toJSString "/abc"
, J.JSString $ J.toJSString "abc/"
, J.JSString $ J.toJSString "../."
, J.JSString $ J.toJSString "g/snapshot"
, J.JSString $ J.toJSString "g/a_mimagex"
, J.JSString $ J.toJSString "g/r;3"
]
-- | Test that the serialisation of 'DiskLogicalId', which is
-- implemented manually, is idempotent. Since we don't have a
-- standalone JSON instance for DiskLogicalId (it's a data type that
-- expands over two fields in a JSObject), we test this by actially
-- testing entire Disk serialisations. So this tests two things at
-- once, basically.
prop_Disk_serialisation :: Disk -> Property
prop_Disk_serialisation = testSerialisation
prop_Disk_array_serialisation :: Disk -> Property
prop_Disk_array_serialisation = testArraySerialisation
-- | Check that node serialisation is idempotent.
prop_Node_serialisation :: Node -> Property
prop_Node_serialisation = testSerialisation
-- | Check that instance serialisation is idempotent.
prop_Inst_serialisation :: Instance -> Property
prop_Inst_serialisation = testSerialisation
-- | Check that address pool serialisation is idempotent.
prop_AddressPool_serialisation :: AddressPool -> Property
prop_AddressPool_serialisation = testSerialisation
-- | Check that network serialisation is idempotent.
prop_Network_serialisation :: Network -> Property
prop_Network_serialisation = testSerialisation
-- | Check that filter action serialisation is idempotent.
prop_FilterAction_serialisation :: FilterAction -> Property
prop_FilterAction_serialisation = testSerialisation
-- | Check that filter predicate serialisation is idempotent.
prop_FilterPredicate_serialisation :: FilterPredicate -> Property
prop_FilterPredicate_serialisation = testSerialisation
-- | Check config serialisation.
prop_Config_serialisation :: Property
prop_Config_serialisation =
forAll (choose (0, maxNodes `div` 4) >>= genEmptyCluster) testSerialisation
-- | Custom HUnit test to check the correspondence between Haskell-generated
-- networks and their Python decoded, validated and re-encoded version.
-- For the technical background of this unit test, check the documentation
-- of "case_py_compat_types" of test/hs/Test/Ganeti/Opcodes.hs
casePyCompatNetworks :: HUnit.Assertion
casePyCompatNetworks = do
let num_networks = 500::Int
networks <- genSample (vectorOf num_networks genValidNetwork)
let networks_with_properties = map getNetworkProperties networks
serialized = J.encode networks
-- check for non-ASCII fields, usually due to 'arbitrary :: String'
mapM_ (\net -> when (any (not . isAscii) (J.encode net)) .
HUnit.assertFailure $
"Network has non-ASCII fields: " ++ show net
) networks
py_stdout <-
runPython "from ganeti import network\n\
\from ganeti import objects\n\
\from ganeti import serializer\n\
\import sys\n\
\net_data = serializer.Load(sys.stdin.read())\n\
\decoded = [objects.Network.FromDict(n) for n in net_data]\n\
\encoded = []\n\
\for net in decoded:\n\
\ a = network.AddressPool(net)\n\
\ encoded.append((a.GetFreeCount(), a.GetReservedCount(), \\\n\
\ net.ToDict()))\n\
\print serializer.Dump(encoded)" serialized
>>= checkPythonResult
let deserialised = J.decode py_stdout::J.Result [(Int, Int, Network)]
decoded <- case deserialised of
J.Ok ops -> return ops
J.Error msg ->
HUnit.assertFailure ("Unable to decode networks: " ++ msg)
-- this already raised an expection, but we need it
-- for proper types
>> fail "Unable to decode networks"
HUnit.assertEqual "Mismatch in number of returned networks"
(length decoded) (length networks_with_properties)
mapM_ (uncurry (HUnit.assertEqual "Different result after encoding/decoding")
) $ zip networks_with_properties decoded
-- | Creates a tuple of the given network combined with some of its properties
-- to be compared against the same properties generated by the python code.
getNetworkProperties :: Network -> (Int, Int, Network)
getNetworkProperties net =
(getFreeCount net, getReservedCount net, net)
-- | Tests the compatibility between Haskell-serialized node groups and their
-- python-decoded and encoded version.
casePyCompatNodegroups :: HUnit.Assertion
casePyCompatNodegroups = do
let num_groups = 500::Int
groups <- genSample (vectorOf num_groups genNodeGroup)
let serialized = J.encode groups
-- check for non-ASCII fields, usually due to 'arbitrary :: String'
mapM_ (\group -> when (any (not . isAscii) (J.encode group)) .
HUnit.assertFailure $
"Node group has non-ASCII fields: " ++ show group
) groups
py_stdout <-
runPython "from ganeti import objects\n\
\from ganeti import serializer\n\
\import sys\n\
\group_data = serializer.Load(sys.stdin.read())\n\
\decoded = [objects.NodeGroup.FromDict(g) for g in group_data]\n\
\encoded = [g.ToDict() for g in decoded]\n\
\print serializer.Dump(encoded)" serialized
>>= checkPythonResult
let deserialised = J.decode py_stdout::J.Result [NodeGroup]
decoded <- case deserialised of
J.Ok ops -> return ops
J.Error msg ->
HUnit.assertFailure ("Unable to decode node groups: " ++ msg)
-- this already raised an expection, but we need it
-- for proper types
>> fail "Unable to decode node groups"
HUnit.assertEqual "Mismatch in number of returned node groups"
(length decoded) (length groups)
mapM_ (uncurry (HUnit.assertEqual "Different result after encoding/decoding")
) $ zip groups decoded
-- | Generates a node group with up to 3 networks.
-- | FIXME: This generates still somewhat completely random data, without normal
-- validation rules.
genNodeGroup :: Gen NodeGroup
genNodeGroup = do
name <- genFQDN
members <- pure []
ndparams <- arbitrary
alloc_policy <- arbitrary
ipolicy <- arbitrary
diskparams <- pure (GenericContainer Map.empty)
num_networks <- choose (0, 3)
net_uuid_list <- vectorOf num_networks (arbitrary::Gen String)
nic_param_list <- vectorOf num_networks (arbitrary::Gen PartialNicParams)
net_map <- pure (GenericContainer . Map.fromList $
zip net_uuid_list nic_param_list)
-- timestamp fields
ctime <- arbitrary
mtime <- arbitrary
uuid <- genFQDN `suchThat` (/= name)
serial <- arbitrary
tags <- Set.fromList <$> genTags
let group = NodeGroup name members ndparams alloc_policy ipolicy diskparams
net_map ctime mtime uuid serial tags
return group
instance Arbitrary NodeGroup where
arbitrary = genNodeGroup
instance Arbitrary Ip4Address where
arbitrary = liftM mkIp4Address $ (,,,) <$> choose (0, 255)
<*> choose (0, 255)
<*> choose (0, 255)
<*> choose (0, 255)
$(genArbitrary ''Ip4Network)
-- | Tests conversions of ip addresses from/to numbers.
prop_ip4AddressAsNum :: Ip4Address -> Property
prop_ip4AddressAsNum ip4 =
ip4AddressFromNumber (ip4AddressToNumber ip4) ==? ip4
-- | Tests that the number produced by 'ip4AddressToNumber' has the correct
-- order of bytes.
prop_ip4AddressToNumber :: Word32 -> Property
prop_ip4AddressToNumber w =
let byte :: Int -> Word32
byte i = (w `div` (256^i)) `mod` 256
ipaddr = List.intercalate "." $ map (show . byte) [3,2..0]
in ip4AddressToNumber <$> readIp4Address ipaddr
==? (return (toInteger w) :: Either String Integer)
-- | IsString instance for 'Ip4Address', to help write the tests.
instance IsString Ip4Address where
fromString s =
fromMaybe (error $ "Failed to parse address from " ++ s) (readIp4Address s)
-- | Tests a few simple cases of IPv4 next address.
caseNextIp4Address :: HUnit.Assertion
caseNextIp4Address = do
HUnit.assertEqual "" "0.0.0.1" $ nextIp4Address "0.0.0.0"
HUnit.assertEqual "" "0.0.0.0" $ nextIp4Address "255.255.255.255"
HUnit.assertEqual "" "1.2.3.5" $ nextIp4Address "1.2.3.4"
HUnit.assertEqual "" "1.3.0.0" $ nextIp4Address "1.2.255.255"
HUnit.assertEqual "" "1.2.255.63" $ nextIp4Address "1.2.255.62"
-- | Tests the compatibility between Haskell-serialized instances and their
-- python-decoded and encoded version.
-- Note: this can be enhanced with logical validations on the decoded objects
casePyCompatInstances :: HUnit.Assertion
casePyCompatInstances = do
let num_inst = 500::Int
instances <- genSample (vectorOf num_inst genInst)
let serialized = J.encode instances
-- check for non-ASCII fields, usually due to 'arbitrary :: String'
mapM_ (\inst -> when (any (not . isAscii) (J.encode inst)) .
HUnit.assertFailure $
"Instance has non-ASCII fields: " ++ show inst
) instances
py_stdout <-
runPython "from ganeti import objects\n\
\from ganeti import serializer\n\
\import sys\n\
\inst_data = serializer.Load(sys.stdin.read())\n\
\decoded = [objects.Instance.FromDict(i) for i in inst_data]\n\
\encoded = [i.ToDict() for i in decoded]\n\
\print serializer.Dump(encoded)" serialized
>>= checkPythonResult
let deserialised = J.decode py_stdout::J.Result [Instance]
decoded <- case deserialised of
J.Ok ops -> return ops
J.Error msg ->
HUnit.assertFailure ("Unable to decode instance: " ++ msg)
-- this already raised an expection, but we need it
-- for proper types
>> fail "Unable to decode instances"
HUnit.assertEqual "Mismatch in number of returned instances"
(length decoded) (length instances)
mapM_ (uncurry (HUnit.assertEqual "Different result after encoding/decoding")
) $ zip instances decoded
-- | A helper function for creating 'LIDPlain' values.
mkLIDPlain :: String -> String -> DiskLogicalId
mkLIDPlain = (LIDPlain .) . LogicalVolume
-- | Tests that the logical ID is correctly found in a plain disk
caseIncludeLogicalIdPlain :: HUnit.Assertion
caseIncludeLogicalIdPlain =
let vg_name = "xenvg" :: String
lv_name = "1234sdf-qwef-2134-asff-asd2-23145d.data" :: String
lv = LogicalVolume vg_name lv_name
time = TOD 0 0
d = RealDisk $
RealDiskData (LIDPlain lv) [] ["node1.example.com"] "diskname"
1000 DiskRdWr
Nothing Nothing Nothing "asdfgr-1234-5123-daf3-sdfw-134f43"
0 time time
in
HUnit.assertBool "Unable to detect that plain Disk includes logical ID" $
includesLogicalId lv d
-- | Tests that the logical ID is correctly found in a DRBD disk
caseIncludeLogicalIdDrbd :: HUnit.Assertion
caseIncludeLogicalIdDrbd =
let vg_name = "xenvg" :: String
lv_name = "1234sdf-qwef-2134-asff-asd2-23145d.data" :: String
time = TOD 0 0
d = RealDisk $
RealDiskData
(LIDDrbd8 "node1.example.com" "node2.example.com" 2000 1 5 "secret")
[ RealDisk $ RealDiskData (mkLIDPlain "onevg" "onelv") []
["node1.example.com", "node2.example.com"] "disk1" 1000 DiskRdWr
Nothing Nothing Nothing "145145-asdf-sdf2-2134-asfd-534g2x"
0 time time
, RealDisk $ RealDiskData (mkLIDPlain vg_name lv_name) []
["node1.example.com", "node2.example.com"] "disk2" 1000 DiskRdWr
Nothing Nothing Nothing "6gd3sd-423f-ag2j-563b-dg34-gj3fse"
0 time time
] ["node1.example.com", "node2.example.com"] "diskname" 1000 DiskRdWr
Nothing Nothing Nothing
"asdfgr-1234-5123-daf3-sdfw-134f43" 0 time time
in
HUnit.assertBool "Unable to detect that plain Disk includes logical ID" $
includesLogicalId (LogicalVolume vg_name lv_name) d
-- | Tests that the logical ID is correctly NOT found in a plain disk
caseNotIncludeLogicalIdPlain :: HUnit.Assertion
caseNotIncludeLogicalIdPlain =
let vg_name = "xenvg" :: String
lv_name = "1234sdf-qwef-2134-asff-asd2-23145d.data" :: String
time = TOD 0 0
d = RealDisk $
RealDiskData (mkLIDPlain "othervg" "otherlv") [] ["node1.example.com"]
"diskname" 1000 DiskRdWr Nothing Nothing Nothing
"asdfgr-1234-5123-daf3-sdfw-134f43"
0 time time
in
HUnit.assertBool "Unable to detect that plain Disk includes logical ID" $
not (includesLogicalId (LogicalVolume vg_name lv_name) d)
testSuite "Objects"
[ 'prop_fillDict
, 'prop_LogicalVolume_serialisation
, 'prop_LogicalVolume_deserialisationFail
, 'prop_Disk_serialisation
, 'prop_Disk_array_serialisation
, 'prop_Inst_serialisation
, 'prop_AddressPool_serialisation
, 'prop_Network_serialisation
, 'prop_Node_serialisation
, 'prop_Config_serialisation
, 'prop_FilterAction_serialisation
, 'prop_FilterPredicate_serialisation
, 'casePyCompatNetworks
, 'casePyCompatNodegroups
, 'casePyCompatInstances
, 'prop_ip4AddressAsNum
, 'prop_ip4AddressToNumber
, 'caseNextIp4Address
, 'caseIncludeLogicalIdPlain
, 'caseIncludeLogicalIdDrbd
, 'caseNotIncludeLogicalIdPlain
]
| grnet/snf-ganeti | test/hs/Test/Ganeti/Objects.hs | bsd-2-clause | 29,828 | 0 | 25 | 7,080 | 5,677 | 2,928 | 2,749 | 535 | 2 |
import Test.Hspec
import Test.QuickCheck
import Control.Exception (evaluate)
import GameOfLife
main :: IO ()
main = hspec $ do
describe "GameOfLife" $ do
describe "NextGeneration" $ do
it "Create valid next glider" $ do
nextGeneration (glider 5 5) `shouldBe` [
[False,False,False,False,False],
[True,False,True,False,False],
[False,True,True,False,False],
[False,True,False,False,False],
[False,False,False,False,False]]
it "Should kill all" $ do
nextGeneration (gameGrid (2, 2) [(1,1),(2,1)]) `shouldBe` [[False, False],[False, False]]
it "Should born new cell" $ do
nextGeneration (gameGrid (3, 3) [(1,1),(2,1), (1,2)]) `shouldBe` [[True,True,False],[True,True,False],[False,False,False]]
describe "Show grid" $ do
it "Should render right" $ do
showGrid (glider 5 5) `shouldBe` "-@---\n--@--\n@@@--\n-----\n-----\n"
it "Should not fail on empty" $ do
showGrid ([]) `shouldBe` ""
| AM636E/HaskellGameOfLife | test/Spec.hs | bsd-3-clause | 1,165 | 0 | 22 | 387 | 421 | 239 | 182 | 24 | 1 |
-- jrc -- json-rpc request client
import qualified Data.Aeson as Aeson
import qualified Data.ByteString.Lazy.Char8 as B8L
import qualified Data.Streaming.Network as NetStream
import qualified Options.Applicative as Opts
import qualified Network.Socket as Socket
import Options.Applicative ((<>))
import qualified JsonRpc.Request as Request
import qualified JsonRpc.Response as Response
data Options = Options
{ optSocket :: String
, optMethod :: String
} deriving (Show)
readCliOpts :: IO Options
readCliOpts =
Opts.execParser $ Opts.info (Opts.helper <*> cliOpts)
( Opts.fullDesc
<> Opts.header "jrc -- JSON-RPC request client"
<> Opts.progDesc "Query a JSON-RPC server." )
where
cliOpts = Options
<$> Opts.argument Opts.str
( Opts.metavar "SOCKET"
<> Opts.help "Path to socket file." )
<*> Opts.argument Opts.str
( Opts.metavar "METHOD"
<> Opts.help "Method name to call." )
main :: IO ()
main = readCliOpts >>= \o ->
let
socketPath = optSocket o
method = optMethod o
request = Request.request method
msgData = Aeson.encode request
in do
sock <- NetStream.getSocketUnix socketPath
Socket.send sock $ B8L.unpack msgData
responseData <- Socket.recv sock 4096
Socket.close sock
let result = case Aeson.decode (B8L.pack responseData) of
Just x -> Response.result x
Nothing -> "Error parsing response: " ++ responseData
putStrLn result
| shmookey/bc-tools | src/jrc.hs | bsd-3-clause | 1,515 | 0 | 18 | 367 | 399 | 211 | 188 | 41 | 2 |
{-# LANGUAGE TypeOperators #-}
-- | Test of long-life for dynamics, via three rotations:
-- rotate 5 dynamic behaviors
-- rotate 6 inputs on left
-- rotate 7 inputs on right
-- for a total 210 combinations based on clock.
module Main
( main
) where
import Sirea.Prelude
import Sirea.Clock
import Sirea.Time
rotate567 :: (Partition p) => B (S p ()) (S p Int :&: S p Int :&: S p Int)
rotate567 = bclockOfFreq 3 >>> bfmap tmNanos >>> bfmap (`div` 333333333) >>>
(bfmap (`mod` 5) &&& bfmap (`mod` 6) &&& bfmap (`mod` 7)) >>>
(bfmap fromInteger *** bfmap fromInteger *** bfmap fromInteger)
add, sub, mul, pow, ssq :: B (S p Int :&: S p Int) (S p Int)
add = bzipWith (+)
sub = bzipWith (-)
pow = bzipWith (^)
mul = bzipWith (*)
ssq = bzipWith (\ b c -> (b*b) + (c*c))
nToF :: Int -> B (S p Int :&: S p Int) (S p Int)
nToF 0 = add
nToF 1 = sub
nToF 2 = pow
nToF 3 = mul
nToF 4 = ssq
nToF _ = error "illegal behavior"
nToFEval :: B (S p Int :&: S p Int :&: S p Int) (S p Int)
nToFEval = bfirst (bfmap nToF) >>> bbeval 0 >>> bright (bconst 999) >>> bmerge
zipAll :: B (S p a :&: S p b :&: S p c) (S p (a,(b,c)))
zipAll = bsecond bzip >>> bzip
main :: IO ()
main = runSireaApp $
rotate567 >>>
(zipAll &&& nToFEval) >>>
bzip >>>
bprint
| dmbarbour/Sirea | tst/RotDyn.hs | bsd-3-clause | 1,298 | 0 | 11 | 332 | 576 | 307 | 269 | 33 | 1 |
{-# LANGUAGE OverloadedStrings #-}
module Database.Tempodb.Methods.Series
( seriesCreate
, seriesGet
, seriesList
, seriesUpdate
, seriesDelete
)
where
import Control.Monad.Reader
import Data.Aeson as A
import Data.ByteString.Char8 as C8
import Data.ByteString.Lazy (fromStrict, toStrict)
import Data.Monoid
import Database.Tempodb.Types
import Database.Tempodb.Util
import Network.HTTP.Base (urlEncodeVars)
import Network.Http.Client
-- | Top-level API methods are:
--
-- 1. Series
-- 2. Read
-- 3. Write
-- 4. Increment
-- 5. Single
-- 6. Delete
--
-- TODO: Docs...
seriesCreate :: ByteString -> Tempodb (Maybe Series)
seriesCreate k = do
let postdata = "{\"key\": \"" <> k <> "\"}"
auth <- ask
req <- liftIO . buildRequest $ do
http POST $ rootpath <> "/series"
setContentLength . fromIntegral $ C8.length postdata
auth
(_,result) <- liftIO . runRequest req $ Just postdata
return . A.decode $ fromStrict result
seriesGet :: IdOrKey -> Tempodb (Maybe Series)
seriesGet q = do
auth <- ask
req <- liftIO . buildRequest $ do
http GET path
auth
(_,result) <- liftIO $ runRequest req Nothing
return . A.decode $ fromStrict result
where
ident (SeriesId i) = "/id/" <> i
ident (SeriesKey k) = "/key/" <> k
path = rootpath <> "/series" <> (ident q)
seriesList :: Maybe QueryArgs -> Tempodb (Maybe [Series])
seriesList q = do
req <- seriesCommon q GET
(_,result) <- liftIO $ runRequest req Nothing
return . A.decode $ fromStrict result
seriesDelete :: Maybe QueryArgs -> Tempodb (Int,ByteString)
seriesDelete q = do
req <- seriesCommon q DELETE
liftIO $ runRequest req Nothing
seriesCommon :: Maybe QueryArgs -> Method -> Tempodb Request
seriesCommon q method = do
auth <- ask
liftIO . buildRequest $ do
http method path
setContentLength 0
auth
where
root = rootpath <> "/series"
path = case q of
Nothing -> root
Just qry -> root <> "?" <> (C8.pack $ urlEncodeVars qry)
seriesUpdate :: IdOrKey -> Series -> Tempodb (Maybe Series)
seriesUpdate q s = do
let postdata = toStrict $ A.encode s
auth <- ask
req <- liftIO . buildRequest $ do
http PUT path
setContentLength . fromIntegral $ C8.length postdata
auth
(_,result) <- liftIO . runRequest req $ Just postdata
return . A.decode $ fromStrict result
where
ident (SeriesId i) = "/id/" <> i
ident (SeriesKey k) = "/key/" <> k
path = rootpath <> "/series" <> (ident q)
| ixmatus/hs-tempodb | src/Database/Tempodb/Methods/Series.hs | bsd-3-clause | 2,668 | 0 | 13 | 733 | 853 | 427 | 426 | 70 | 2 |
{-# LANGUAGE CPP #-}
{-# LANGUAGE OverloadedStrings #-}
module Dhall.Test.TypeInference where
import Control.Exception (SomeException (..))
import Data.Text (Text)
import Prelude hiding (FilePath)
import Test.Tasty (TestTree)
import Turtle (FilePath, (</>))
import qualified Control.Exception as Exception
import qualified Control.Monad as Monad
import qualified Data.Text as Text
import qualified Data.Text.IO as Text.IO
import qualified Dhall.Core as Core
import qualified Dhall.Import as Import
import qualified Dhall.Parser as Parser
import qualified Dhall.Test.Util as Test.Util
import qualified Dhall.TypeCheck as TypeCheck
import qualified System.FilePath as FilePath
import qualified Test.Tasty as Tasty
import qualified Test.Tasty.HUnit as Tasty.HUnit
import qualified Turtle
typeInferenceDirectory :: FilePath
typeInferenceDirectory = "./dhall-lang/tests/type-inference"
getTests :: IO TestTree
getTests = do
let successTestFiles = do
path <- Turtle.lstree (typeInferenceDirectory </> "success")
let skip =
-- These tests intermittently fails with:
-- "Error: Remote host not found"
[ typeInferenceDirectory </> "success/CacheImportsA.dhall"
, typeInferenceDirectory </> "success/CacheImportsCanonicalizeA.dhall"
]
Monad.guard (path `notElem` skip)
return path
successTests <- Test.Util.discover (Turtle.chars <* "A.dhall") successTest successTestFiles
let failureTestFiles = Turtle.lstree (typeInferenceDirectory </> "failure")
failureTests <- Test.Util.discover (Turtle.chars <* ".dhall") failureTest failureTestFiles
let testTree = Tasty.testGroup "type-inference tests"
[ successTests
, failureTests
]
return testTree
successTest :: Text -> TestTree
successTest prefix = do
let expectedFailures =
[]
#ifdef WITH_HTTP
#else
++ [ typeInferenceDirectory </> "success/CacheImports"
]
#endif
Test.Util.testCase prefix expectedFailures $ do
let prefixFP = Text.unpack prefix
actualCode <- Text.IO.readFile (prefixFP <> "A.dhall")
actualExpr <- Core.throws (Parser.exprFromText mempty actualCode)
tryResolvedExpr <-
Exception.try
(Test.Util.loadRelativeTo
(FilePath.takeDirectory prefixFP)
Import.IgnoreSemanticCache
(Core.denote actualExpr))
resolvedExpr <- case tryResolvedExpr of
Left exception -> Tasty.HUnit.assertFailure (show (exception :: SomeException))
Right resolvedExpr -> return resolvedExpr
expectedTypeCode <- Text.IO.readFile (prefixFP <> "B.dhall")
expectedType <- Core.throws (Parser.exprFromText mempty expectedTypeCode)
resolvedExpectedType <- Import.assertNoImports (Core.denote expectedType)
inferredType <- case TypeCheck.typeOf resolvedExpr of
Left exception -> Tasty.HUnit.assertFailure (show exception)
Right inferredType -> return inferredType
let message = "The inferred type did not match the expected type"
Tasty.HUnit.assertEqual message resolvedExpectedType inferredType
-- We also add this to exercise the `Dhall.Eval.conv` code path, since
-- it's easy to forget to update it when adding new syntax
_ <- Core.throws (TypeCheck.typeOf (Core.Annot resolvedExpr resolvedExpectedType))
return ()
failureTest :: Text -> TestTree
failureTest prefix = do
let expectedFailures =
[
-- Duplicate fields are incorrectly caught during parsing:
-- https://github.com/dhall-lang/dhall-haskell/issues/772
typeInferenceDirectory </> "failure/unit/RecordTypeDuplicateFields"
, typeInferenceDirectory </> "failure/unit/UnionTypeDuplicateVariants1"
, typeInferenceDirectory </> "failure/unit/UnionTypeDuplicateVariants2"
]
Test.Util.testCase prefix expectedFailures $ do
let prefixFP = Text.unpack prefix
code <- Text.IO.readFile (prefixFP <> ".dhall")
expression <- case Parser.exprFromText mempty code of
Left _ -> Tasty.HUnit.assertFailure (prefixFP <> " should have parsed")
Right e -> return e
resolved <- Import.assertNoImports expression
case TypeCheck.typeOf resolved of
Left _ -> return ()
Right _ -> Tasty.HUnit.assertFailure (prefixFP <> " should not have type-checked")
| Gabriel439/Haskell-Dhall-Library | dhall/tests/Dhall/Test/TypeInference.hs | bsd-3-clause | 4,759 | 0 | 17 | 1,288 | 952 | 500 | 452 | 84 | 3 |
{-# LANGUAGE RankNTypes #-}
-- This program must be called with GHC's libdir as the single command line
-- argument.
module Main where
-- import Data.Generics
import Data.Data
import Data.List
import System.IO
import GHC
import BasicTypes
import DynFlags
import MonadUtils
import Outputable
import Bag (filterBag,isEmptyBag)
import System.Directory (removeFile)
import System.Environment( getArgs )
import qualified Data.Map as Map
import Data.Dynamic ( fromDynamic,Dynamic )
main::IO()
main = do
[libdir] <- getArgs
testOneFile libdir "AnnotationTuple"
testOneFile libdir fileName = do
((anns,cs),p) <- runGhc (Just libdir) $ do
dflags <- getSessionDynFlags
setSessionDynFlags dflags
let mn =mkModuleName fileName
addTarget Target { targetId = TargetModule mn
, targetAllowObjCode = True
, targetContents = Nothing }
load LoadAllTargets
modSum <- getModSummary mn
p <- parseModule modSum
t <- typecheckModule p
d <- desugarModule t
l <- loadModule d
let ts=typecheckedSource l
r =renamedSource l
return (pm_annotations p,p)
let tupArgs = gq (pm_parsed_source p)
putStrLn (pp tupArgs)
putStrLn (intercalate "\n" [showAnns anns])
where
gq ast = everything (++) ([] `mkQ` doLHsTupArg) ast
doLHsTupArg :: LHsTupArg GhcPs -> [(SrcSpan,String,HsExpr GhcPs)]
doLHsTupArg (L l arg@(Present {}))
= [(l,"p",ExplicitTuple noExtField [L l arg] Boxed)]
doLHsTupArg (L l arg@(Missing {}))
= [(l,"m",ExplicitTuple noExtField [L l arg] Boxed)]
showAnns anns = "[\n" ++ (intercalate "\n"
$ map (\((s,k),v)
-> ("(AK " ++ pp s ++ " " ++ show k ++" = " ++ pp v ++ ")\n"))
$ Map.toList anns)
++ "]\n"
pp a = showPpr unsafeGlobalDynFlags a
-- ---------------------------------------------------------------------
-- Copied from syb for the test
-- | Generic queries of type \"r\",
-- i.e., take any \"a\" and return an \"r\"
--
type GenericQ r = forall a. Data a => a -> r
-- | Make a generic query;
-- start from a type-specific case;
-- return a constant otherwise
--
mkQ :: ( Typeable a
, Typeable b
)
=> r
-> (b -> r)
-> a
-> r
(r `mkQ` br) a = case cast a of
Just b -> br b
Nothing -> r
-- | Summarise all nodes in top-down, left-to-right order
everything :: (r -> r -> r) -> GenericQ r -> GenericQ r
-- Apply f to x to summarise top-level node;
-- use gmapQ to recurse into immediate subterms;
-- use ordinary foldl to reduce list of intermediate results
everything k f x = foldl k (f x) (gmapQ (everything k f) x)
| sdiehl/ghc | testsuite/tests/ghc-api/annotations/parseTree.hs | bsd-3-clause | 3,062 | 1 | 20 | 1,053 | 820 | 430 | 390 | 63 | 2 |
{-# LANGUAGE FlexibleInstances, ScopedTypeVariables, TypeSynonymInstances #-}
-- |This module contains classes and functions to relate Haskell types
-- with OpenGL DataTypes (typically used to describe the values stored
-- in arrays) and VariableTypes (used as attributes and uniforms in
-- GLSL programs).
module GLUtil.TypeMapping where
import Data.Int
import Data.Word
import Foreign.Storable (Storable)
import qualified Graphics.Rendering.OpenGL as GL
import Graphics.Rendering.OpenGL
import Linear (V1, V2, V3, V4, M22, M33, M44)
-- | A mapping from Haskell types to values of 'VariableType'. This
-- defines how Haskell values may be mapped to values that may be
-- bound to GLSL variables.
class HasVariableType a where
variableType :: a -> VariableType
instance HasVariableType Float where variableType _ = Float'
instance HasVariableType Int32 where variableType _ = Int'
instance HasVariableType Word32 where variableType _ = UnsignedInt'
instance HasVariableType (V1 GLfloat) where variableType _ = Float'
instance HasVariableType (V2 GLfloat) where variableType _ = FloatVec2
instance HasVariableType (V3 GLfloat) where variableType _ = FloatVec3
instance HasVariableType (V4 GLfloat) where variableType _ = FloatVec4
instance HasVariableType (V1 Int32) where variableType _ = Int'
instance HasVariableType (V2 Int32) where variableType _ = IntVec2
instance HasVariableType (V3 Int32) where variableType _ = IntVec3
instance HasVariableType (V4 Int32) where variableType _ = IntVec4
instance HasVariableType (V1 Word32) where variableType _ = UnsignedInt'
instance HasVariableType (V2 Word32) where variableType _ = UnsignedIntVec2
instance HasVariableType (V3 Word32) where variableType _ = UnsignedIntVec3
instance HasVariableType (V4 Word32) where variableType _ = UnsignedIntVec4
instance HasVariableType (M22 GLfloat) where variableType _ = FloatMat2
instance HasVariableType (M33 GLfloat) where variableType _ = FloatMat3
instance HasVariableType (M44 GLfloat) where variableType _ = FloatMat4
instance forall t. HasVariableType t => HasVariableType [t] where
variableType _ = variableType (undefined::t)
-- | Maps each 'VariableType' to its corresponding
-- 'DataType'. Typically this indicates the element type of composite
-- variable types (e.g. @variableDataType FloatVec2 = Float@). Note
-- that this is a partial mapping as we are primarily supporting the
-- use of these types as inputs to GLSL programs where types such as
-- Bool are not supported.
variableDataType :: VariableType -> DataType
variableDataType Float' = GL.Float
variableDataType FloatVec2 = GL.Float
variableDataType FloatVec3 = GL.Float
variableDataType FloatVec4 = GL.Float
variableDataType Int' = GL.Int
variableDataType IntVec2 = GL.Int
variableDataType IntVec3 = GL.Int
variableDataType IntVec4 = GL.Int
variableDataType UnsignedInt' = GL.UnsignedInt
variableDataType UnsignedIntVec2 = GL.UnsignedInt
variableDataType UnsignedIntVec3 = GL.UnsignedInt
variableDataType UnsignedIntVec4 = GL.UnsignedInt
variableDataType FloatMat2 = GL.Float
variableDataType FloatMat3 = GL.Float
variableDataType FloatMat4 = GL.Float
variableDataType FloatMat2x3 = GL.Float
variableDataType FloatMat2x4 = GL.Float
variableDataType FloatMat3x2 = GL.Float
variableDataType FloatMat3x4 = GL.Float
variableDataType FloatMat4x2 = GL.Float
variableDataType FloatMat4x3 = GL.Float
variableDataType _ = error "Unsupported variable type!"
-- |Open mapping from Haskell types to OpenGL types.
class Storable a => HasGLType a where
glType :: a -> DataType
instance HasGLType GLint where glType _ = GL.Int
instance HasGLType Word8 where glType _ = GL.UnsignedByte
instance HasGLType Word16 where glType _ = GL.UnsignedShort
instance HasGLType Word32 where glType _ = GL.UnsignedInt
instance HasGLType Float where glType _ = GL.Float
| coghex/abridgefaraway | src/GLUtil/TypeMapping.hs | bsd-3-clause | 3,818 | 0 | 7 | 512 | 863 | 450 | 413 | 60 | 1 |
{- Provides constants used by the CORDIC algorithm.
- `alistF` is a list of angles [ atan 1, atan (1/2), atan (1/4, ... ]
- `klistF` is a list of the scaling constants for each iteration
- Traditionally these would have been hard-coded for performance; they are
- generated programmatically here for simplicity.
-}
module Util
(
alistF,
klistF
) where
import Numeric.Fixed
-- |Fixed-point cast of alist
alistF = map toFixed alist
-- |Fixed-point cast of klist
klistF = map toFixed klist
-- |Infinite list of angles with tangent ratios [1, 1/(2^i)]
alist :: [Double]
alist = [ atan ( 1 / 2 ^ e ) | e <- [ 0 .. ] ]
-- |Infinite list of scaling factors
klist :: [Double]
klist = klist' 1 ( k 0 )
-- |Recursive generator for scaling factors
klist' :: Int -> Double -> [Double]
klist' i n = n : klist' ( i + 1 ) ( k i * n )
-- |Scaling factor k at iteration i
k :: Int -> Double
k i = 1 / sqrt ( 1 + 2 ^^ (( -2 ) * i ))
| chronon-io/cordic | src/Util.hs | bsd-3-clause | 935 | 0 | 12 | 212 | 209 | 117 | 92 | 15 | 1 |
-- | RSS 2.0 feed source.
-- Items are produced from RSS 2.0 feeds.
-- The itmName, itmURL and itmDescription fields are filled from the
-- corresponding RSS fields.
-- The itmTags field is populated from the RSS tags.
--
-- TODO: support for other feed formats.
module Network.LambNyaa.Source.Feed (rssFeed, rssFeed') where
import Network.LambNyaa.Item
import Network.LambNyaa.Monad
import Network.LambNyaa.Log
import Text.RSS.Syntax
import Text.Feed.Types hiding (Item)
import Network.Download
import Control.Monad
warn' = warn "Source.Feed"
err' = err "Source.Feed"
-- | Create a Source from an RSS feed. The itmSource field of Items originating
-- from this Source will contain the URL of the feed.
rssFeed :: URL -> Nyaa [Item]
rssFeed url = rssFeed' url url
-- | Create a Source from a named RSS feed. The itmSource field of Items
-- originating from this Source will contain the given name.
rssFeed' :: String -> URL -> Nyaa [Item]
rssFeed' src url = source $ do
ef <- openAsFeed url
case ef of
Right (RSSFeed rss) -> do
let is = getItems src (rssChannel rss)
when (null is) . warn' $ "No RSS items from feed " ++ url ++ "!"
return is
_ -> do
err' $ "Unable to parse RSS feed from " ++ url ++ "!"
return []
getItems :: String -> RSSChannel -> [Item]
getItems src = map (getItem src) . rssItems
getItem :: String -> RSSItem -> Item
getItem src item = Item {
itmIdentifier = undefined,
itmSeenBefore = undefined,
itmName = maybe "" id (rssItemTitle item),
itmURL = maybe "" id (rssItemLink item),
itmDescription = rssItemDescription item,
itmSource = src,
itmTags = map rssCategoryValue $ rssItemCategories item
}
| valderman/lambnyaa | Network/LambNyaa/Source/Feed.hs | bsd-3-clause | 1,739 | 0 | 19 | 392 | 408 | 219 | 189 | 34 | 2 |
module Data.TestPolynomial where
import Prelude hiding (Rational, gcd, negate, (*), (+), (-), (/), (^))
import Domains.Euclidean
import Domains.Ring
import Data.Polynomial
import Data.Rational
import Collins
import TestUtil
x, x2, x3, x4, x5, x6, x7, x8 :: Polynomial Integer
x = Term 1 1 (Data.Polynomial.Const 0)
x2 = x * x
x3 = x2 * x
x4 = x3 * x
x5 = x4 * x
x6 = x5 * x
x7 = x6 * x
x8 = x7 * x
p :: Polynomial Integer
p = Term 1 1 (Data.Polynomial.Const 1)
pc :: Integer -> Polynomial Integer
pc = promote
run :: IO ()
run
-- print (Collins.gcd (pc 0) (pc 1))
-- print (Collins.gcd (pc 1) (pc 0))
-- print (Collins.gcd (pc 3) (pc 7))
-- print (Collins.gcd (pc 3 * x + pc 1) (pc 7))
= do
test "gcd" (x - pc 1) (gcd (x2 - pc 1) (x - pc 1))
putStrLn ("gcd " ++ show (gcd (pc 6 * (x2 - pc 1)) (pc 4 * (x - pc 1))))
test "ratio " (rational (pc 3 * x + pc 3) (pc 2)) (rational (pc 6 * (x2 - pc 1)) (pc 4 * (x - pc 1)))
test "gcd" (x - one) (gcd (x2 - pc 1) (x2 - pc 2 * x + pc 1))
test "ratio " (rational (x + one) (x - one)) (rational (x2 - pc 1) (x2 - pc 2 * x + pc 1))
test "gcd" one (gcd (pc 3 * x2 + pc 1) (pc 5 * x4 + x2 + pc 4))
test
"subresultant"
260708
(Collins.resultant
(x8 + x6 - pc 3 * x4 - pc 3 * x3 + pc 8 * x2 + pc 2 * x - pc 5)
(pc 3 * x6 + pc 5 * x4 - pc 4 * x2 - pc 9 * x + pc 21))
putStrLn ("p = " ++ show (p * p * p))
| pmilne/algebra | test/Data/TestPolynomial.hs | bsd-3-clause | 1,512 | 2 | 20 | 502 | 782 | 407 | 375 | 37 | 1 |
{-# LANGUAGE StandaloneDeriving #-}
module Data.Text.Index
(
Index
, lookupWord
, lookupPhrase
, Weight, toDouble
, size
, empty
, addDocument
, removeDocument
)where
import qualified Data.Search.Results as Results
import Data.Foldable (foldMap, for_)
import Data.Functor ((<$>))
import Data.List (foldl', sortOn, intercalate)
import Data.Monoid ((<>), Endo(Endo, appEndo))
import Data.Ord (Down(Down))
import qualified Data.OrdPSQ as PSQ
import qualified Data.Text as Text
import Data.Text (Text)
import qualified Data.Text.ICU as ICU
import qualified Data.Text.ICU.Char as ICU
import qualified Data.Text.ICU.Normalize as ICU
import qualified Data.TST as TST
import Data.TST (TST)
import Data.Typeable (Typeable)
import GHC.Generics (Generic)
import Prelude hiding (lookup, words, Word)
-- TODO: exact matches within long strings are not found right now (unless
-- the start of the match happens to coincide with a chunk boundary).
-- The index contains many (trimmed-down) variants of all source words.
type Index id
= TST Char (Entries id)
-- For a given variant (this is a leaf of the TST), we have a set
-- of original words, together with the source id, and its weight (indicating
-- how much the original word was trimmed).
type Entries id
= PSQ.OrdPSQ id Weight Word
-- A 'Word' does not contain spaces.
type Word
= Text
newtype Weight
= Weight Double
deriving (Eq, Show, Typeable)
toDouble :: Weight -> Double
toDouble (Weight d) = d
instance Monoid Weight where
mappend (Weight a) (Weight b) = Weight $ a + b
mempty = Weight 0
reweigh :: (Double -> Double) -> Weight -> Weight
reweigh f (Weight d) = Weight (f d)
instance Ord Weight where
compare (Weight a) (Weight b) = compare a b
-- Lookup.
lookupExact :: (Ord id) =>
Index id -> Word -> Results.T id Weight Word
lookupExact index t = case TST.lookup (realChars t) index of
Just entries -> Results.fromPSQ entries
Nothing -> Results.empty
lookupWord :: (Ord id) =>
Index id -> Word -> Results.T id Weight Word
lookupWord index t = Results.union $ map f variations
where
f (t', w) = Results.changeWeight (w <>) $
lookupExact index t'
variations = deletions defaultWeights =<< chop (normalise t)
lookupPhrase :: (Ord id) => Index id -> Text -> Results.T id Weight Word
lookupPhrase index t = case words t of
-- Empty search term gives no results. Do we want a way
-- to return all results?
[] -> Results.empty
terms -> Results.intersection $ map (lookupWord index) terms
size :: Index id -> Int
size = length . TST.toList
-- Creation.
empty :: Index id
empty = TST.empty
addVariant :: (Ord id) => id -> Word -> (Word, Weight) -> Index id -> Index id
addVariant i original (variant, w) = TST.insertWith
(const $ snd . PSQ.alter f i)
(realChars variant)
(PSQ.singleton i w original)
where
f (Just (w', original'))
| w' <= w = ((), Just (w', original'))
f _ = ((), Just (w , original ))
addWord :: (Ord id) => id -> Word -> Index id -> Index id
addWord i t = appEndo . foldMap (Endo . addVariant i t) $
vary defaultWeights t
addDocument :: (Ord id) => id -> Text -> Index id -> Index id
addDocument i = appEndo . foldMap (Endo . addWord i) . words
removeDocument :: (Ord id) => id -> Index id -> Index id
removeDocument i = fmap (PSQ.delete i)
words :: Text -> [Text]
words = map ICU.brkBreak .
filter ((/=) ICU.Uncategorized . ICU.brkStatus) .
ICU.breaks (ICU.breakWord ICU.Current)
normalise :: Text -> Text
normalise = ICU.normalize ICU.NFKD . Text.toCaseFold
weightedLength :: Weights -> Text -> Weight
weightedLength wts = mconcat . map (character wts) . Text.unpack
-- Generate all variants of a word.
-- Some of these contain holes, others are just shortened versions.
vary :: Weights -> Word -> [(Word, Weight)]
vary wts t = shorten wts =<< deletions wts =<< chop =<< [normalise t]
shorten :: Weights -> (Word, Weight) -> [(Word, Weight)]
shorten wts (t, w)
-- Non-trivial deletion variants of a word are not further shortened.
| w > Weight 0.9 = [(t, w)]
| otherwise = (t, w) : concatMap
(\ br -> weigh (ICU.brkPrefix br) ++ weigh (ICU.brkSuffix br))
breaks
where
-- Break up a string in characters.
breaks = ICU.breaks (ICU.breakCharacter ICU.Current) t
weigh st
| Text.null st = []
| otherwise = [(st, w <> _N <> n)]
where
n = reweigh negate $ weightedLength wts st
_N = weightedLength wts t
deletions :: Weights -> (Word, Weight) -> [(Word, Weight)]
deletions wts (t, w) = step (t, w) where
maximalWeight = budget wts $ weightedLength wts t
step :: (Word, Weight) -> [(Word, Weight)]
step (t, w) = go t w 1 where
go t w start
| w > maximalWeight = []
| Text.null t = [(t, w)]
| otherwise = (t, w) : concatMap
(delete w)
[(i, Text.splitAt i t) | i <- [start .. Text.length t]]
delete w (i, (leftC, right)) = go
(Text.append left (Text.cons hole right))
(w <> cost)
(succ i)
where
c = Text.last leftC
left = Text.init leftC
cost = reweigh (continuationCost *) characterCost
continuationCost
| not (Text.null left) && Text.last left == hole = continuation wts
| otherwise = 1
characterCost = character wts c
data Weights
= Weights
{ continuation :: Double
, character :: Char -> Weight
, budget :: Weight -> Weight
}
defaultWeights :: Weights
defaultWeights = Weights
{ continuation = 0.75
, character = \ c -> case ICU.property ICU.GeneralCategory c of
ICU.NonSpacingMark -> Weight 0.2
ICU.EnclosingMark -> Weight 0.2
ICU.CombiningSpacingMark -> Weight 0.2
ICU.OtherPunctuation -> Weight 0.4
_ -> Weight 1
, budget = reweigh $ min 2 . \ l -> 0.5 + l / 5
}
hole :: Char
hole = '\xFFFC'
showHoles :: Text -> Text
showHoles = Text.map (\ c -> if c == hole then '□' else c)
realChars :: Word -> [Char]
realChars = filter (/= hole) . Text.unpack
-- Dealing with long words.
-- Chop up a word in overlapping chunks, of maximal length 15
-- and typical length 10.
chop :: Word -> [(Word, Weight)]
chop t
| Text.length t <= maximalChunk = [(t, Weight 0)]
| otherwise = (Text.take maximalChunk t, Weight 0) :
(map (flip (,) $ Weight 1) . chop' $ Text.drop overlap t)
where
chop' t
| Text.length t <= maximalChunk = [t]
| otherwise = Text.take typicalChunk t : chop' (Text.drop overlap t)
maximalChunk = 3 * overlap
typicalChunk = 2 * overlap
overlap = 5
-- Helper functions and tests.
printVariations :: [(Word, Weight)] -> IO ()
printVariations = mapM_ $ \ (t, w) -> do
putStr (Text.unpack . showHoles $ t)
putStrLn (" " ++ show w)
testDocument :: Text
testDocument = Text.pack "Dit is een hele zin, met allemaal woorden erin. Ook woorden met accent, als café. Daarnaast wat cijfers: 0 123 4.567. En natuurlijk symbolen ☺☹!"
dr :: Text
dr = Text.pack "driehoeksongelijkheid"
| ariep/text-index | src/Data/Text/Index.hs | bsd-3-clause | 7,323 | 0 | 19 | 1,922 | 2,480 | 1,325 | 1,155 | 164 | 5 |
module Callback.MouseMove where
import qualified Data.IORef as R
import Control.Applicative ((<$>))
import qualified Utils as LU
import qualified Graphics.UI.GLFW as GLFW
import qualified AppData as AP
newMouseMoveCallback :: AP.AppDataRef -> GLFW.CursorPosCallback
newMouseMoveCallback appRef = callback
where
callback win x y = do
mp <- LU.windowToLevelCoords (floor x, floor y) <$> R.readIORef appRef
R.modifyIORef appRef (AP.handleMouseMoved mp)
| dan-t/layers | src/Callback/MouseMove.hs | bsd-3-clause | 480 | 0 | 13 | 86 | 133 | 76 | 57 | 11 | 1 |
module Network.XMPP.Address
( XMPPLocal
, localText
, localFromText
, XMPPDomain
, domainText
, domainFromText
, XMPPResource
, resourceText
, resourceFromText
, XMPPAddress(..)
, xmppAddress'
, xmppAddress
, addressToText
, addressBare
, BareJID(..)
, bareJidGet
, bareJidAddress
, bareJidToText
, FullJID(..)
, fullJidGet
, fullJidAddress
, fullJidToText
) where
import Data.Text (Text)
import Data.Aeson
import Data.Aeson.Types (toJSONKeyText)
import Control.Applicative
import Data.Attoparsec.Text
import Text.StringPrep
import Text.StringPrep.Profiles
import Network.XMPP.Utils
newtype XMPPLocal = XMPPLocal { localText :: Text }
deriving (Eq, Ord, Show)
localFromText :: Text -> Maybe XMPPLocal
localFromText t = XMPPLocal <$> runStringPrep nodePrepProfile t
newtype XMPPDomain = XMPPDomain { domainText :: Text }
deriving (Eq, Ord, Show)
instance FromJSON XMPPDomain where
parseJSON = withText "XMPPDomain" $ \t -> case domainFromText t of
Nothing -> fail "XMPPDomain"
Just r -> return r
instance ToJSON XMPPDomain where
toJSON = toJSON . domainText
domainFromText :: Text -> Maybe XMPPDomain
domainFromText t = XMPPDomain <$> runStringPrep xmppNamePrepProfile t
newtype XMPPResource = XMPPResource { resourceText :: Text }
deriving (Eq, Ord, Show)
resourceFromText :: Text -> Maybe XMPPResource
resourceFromText t = XMPPResource <$> runStringPrep resourcePrepProfile t
data XMPPAddress = XMPPAddress { addressLocal :: Maybe XMPPLocal
, addressDomain :: XMPPDomain
, addressResource :: Maybe XMPPResource
}
deriving (Eq, Ord)
instance Show XMPPAddress where
show = show . addressToText
instance FromJSON XMPPAddress where
parseJSON = withText "XMPPAddress" $ \t -> case xmppAddress t of
Left e -> fail e
Right r -> return r
instance FromJSONKey XMPPAddress where
fromJSONKey = FromJSONKeyTextParser $ \k -> case xmppAddress k of
Left e -> fail e
Right r -> return r
instance ToJSON XMPPAddress where
toJSON = toJSON . addressToText
instance ToJSONKey XMPPAddress where
toJSONKey = toJSONKeyText addressToText
nodeProhibited :: [Range]
nodeProhibited =
[ range '\x0022' '\x0023'
, range '\x0026' '\x0027'
, range '\x0027' '\x0028'
, range '\x002F' '\x0030'
, range '\x003A' '\x003B'
, range '\x003C' '\x003D'
, range '\x003E' '\x003F'
, range '\x0040' '\x0041'
]
nodePrepProfile :: StringPrepProfile
nodePrepProfile =
Profile { maps = [b1, b2]
, shouldNormalize = True
, prohibited = [a1, c11, c12, c21, c22, c3, c4, c5, c6, c7, c8, c9, nodeProhibited]
, shouldCheckBidi = True
}
xmppNamePrepProfile :: StringPrepProfile
xmppNamePrepProfile = namePrepProfile False
resourcePrepProfile :: StringPrepProfile
resourcePrepProfile =
Profile { maps = [b1]
, shouldNormalize = True
, prohibited = [a1, c12, c21, c22, c3, c4, c5, c6, c7, c8, c9]
, shouldCheckBidi = True
}
xmppAddress' :: Parser XMPPAddress
xmppAddress' = do
first <- takeTill (\c -> c == '@' || c == '/')
sep <- optional anyChar
case sep of
Just '@' -> do
addressLocal <- Just <$> checkLocal first
addressDomain <- takeTill (== '/') >>= checkDomain
sep2 <- optional anyChar
case sep2 of
Just '/' -> do
addressResource <- Just <$> (takeText >>= checkResource)
return XMPPAddress { .. }
Nothing -> return XMPPAddress { addressResource = Nothing
, ..
}
_ -> error "xmppAddress: impossible second separator"
Just '/' -> do
addressDomain <- checkDomain first
addressResource <- Just <$> (takeText >>= checkResource)
return XMPPAddress { addressLocal = Nothing
, ..
}
Nothing -> do
addressDomain <- checkDomain first
return XMPPAddress { addressLocal = Nothing
, addressResource = Nothing
, ..
}
_ -> error "xmppAddress: impossible first separator"
where checkLocal = maybeFail "xmppAddress: localpart doesn't satisfy Nodeprep profile of stringprep" . localFromText
checkDomain = maybeFail "xmppAddress: domainpart doesn't satisfy Nameprep profile of stringprep" . domainFromText
checkResource = maybeFail "xmppAddress: resourcepart doesn't satisfy Resourceprep profile of stringprep" . resourceFromText
xmppAddress :: Text -> Either String XMPPAddress
xmppAddress = parseValue xmppAddress'
addressToText :: XMPPAddress -> Text
addressToText (XMPPAddress {..}) =
maybe mempty ((<> "@") . localText) addressLocal
<> domainText addressDomain
<> maybe mempty (("/" <>) . resourceText) addressResource
addressBare :: XMPPAddress -> XMPPAddress
addressBare (XMPPAddress {..}) = XMPPAddress { addressResource = Nothing
, ..
}
data BareJID = BareJID { bareLocal :: XMPPLocal
, bareDomain :: XMPPDomain
}
deriving (Eq, Ord)
bareJidAddress :: BareJID -> XMPPAddress
bareJidAddress (BareJID {..}) = XMPPAddress (Just bareLocal) bareDomain Nothing
bareJidToText :: BareJID -> Text
bareJidToText = addressToText . bareJidAddress
instance Show BareJID where
show = show . bareJidToText
instance FromJSON BareJID where
parseJSON v = do
addr <- parseJSON v
case bareJidGet addr of
Nothing -> fail "BareJID"
Just r -> return r
instance ToJSON BareJID where
toJSON = toJSON . bareJidAddress
bareJidGet :: XMPPAddress -> Maybe BareJID
bareJidGet XMPPAddress { addressLocal = Just bareLocal, addressDomain = bareDomain, addressResource = Nothing } = Just BareJID {..}
bareJidGet _ = Nothing
data FullJID = FullJID { fullBare :: BareJID
, fullResource :: XMPPResource
}
deriving (Eq, Ord)
fullJidAddress :: FullJID -> XMPPAddress
fullJidAddress (FullJID {..}) = XMPPAddress (Just $ bareLocal fullBare) (bareDomain fullBare) (Just fullResource)
fullJidToText :: FullJID -> Text
fullJidToText = addressToText . fullJidAddress
instance Show FullJID where
show = show . fullJidToText
instance FromJSON FullJID where
parseJSON v = do
addr <- parseJSON v
case fullJidGet addr of
Nothing -> fail "FullJID"
Just r -> return r
instance ToJSON FullJID where
toJSON = toJSON . fullJidAddress
fullJidGet :: XMPPAddress -> Maybe FullJID
fullJidGet XMPPAddress {..} = do
bareLocal <- addressLocal
fullResource <- addressResource
return FullJID { fullBare = BareJID { bareDomain = addressDomain, .. }, .. }
| abbradar/yaxmpp | src/Network/XMPP/Address.hs | bsd-3-clause | 7,050 | 0 | 20 | 1,968 | 1,789 | 956 | 833 | -1 | -1 |
{-# OPTIONS -fno-implicit-prelude #-}
-----------------------------------------------------------------------------
-- |
-- Module : Foreign.Marshal.Error
-- Copyright : (c) The FFI task force 2001
-- License : BSD-style (see the file libraries/base/LICENSE)
--
-- Maintainer : [email protected]
-- Stability : provisional
-- Portability : portable
--
-- Marshalling support: Handling of common error conditions
--
-----------------------------------------------------------------------------
module Foreign.Marshal.Error (
-- * Error utilities
-- |Throw an exception on specific return values
--
throwIf, -- :: (a -> Bool) -> (a -> String) -> IO a -> IO a
throwIf_, -- :: (a -> Bool) -> (a -> String) -> IO a -> IO ()
throwIfNeg, -- :: (Ord a, Num a)
-- => (a -> String) -> IO a -> IO a
throwIfNeg_, -- :: (Ord a, Num a)
-- => (a -> String) -> IO a -> IO ()
throwIfNull, -- :: String -> IO (Ptr a) -> IO (Ptr a)
-- Discard return value
--
void -- IO a -> IO ()
) where
import Foreign.Ptr
-- exported functions
-- ------------------
-- |Guard an 'IO' operation and throw an exception if the result meets the given
-- predicate
--
-- * the second argument computes an error message from the result of the 'IO'
-- operation
--
throwIf :: (a -> Bool) -> (a -> String) -> IO a -> IO a
throwIf pred msgfct act =
do
res <- act
(if pred res then ioError . userError . msgfct else return) res
-- |Like 'throwIf', but discarding the result
--
throwIf_ :: (a -> Bool) -> (a -> String) -> IO a -> IO ()
throwIf_ pred msgfct act = void $ throwIf pred msgfct act
-- |Guards against negative result values
--
throwIfNeg :: (Ord a, Num a) => (a -> String) -> IO a -> IO a
throwIfNeg = throwIf (< 0)
-- |Like 'throwIfNeg', but discarding the result
--
throwIfNeg_ :: (Ord a, Num a) => (a -> String) -> IO a -> IO ()
throwIfNeg_ = throwIf_ (< 0)
-- |Guards against null pointers
--
throwIfNull :: String -> IO (Ptr a) -> IO (Ptr a)
throwIfNull = throwIf (== nullPtr) . const
-- |Discard the return value of an 'IO' action
--
void :: IO a -> IO ()
void act = act >> return ()
| OS2World/DEV-UTIL-HUGS | libraries/Foreign/Marshal/Error.hs | bsd-3-clause | 2,293 | 0 | 11 | 617 | 412 | 236 | 176 | 24 | 2 |
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE DeriveDataTypeable #-}
{-# LANGUAGE TypeFamilies #-}
-- | Types used while planning how to build everything in a project.
--
-- Primarily this is the 'ElaboratedInstallPlan'.
--
module Distribution.Client.ProjectPlanning.Types (
SolverInstallPlan,
-- * Elaborated install plan types
ElaboratedInstallPlan,
ElaboratedConfiguredPackage(..),
elabDistDirParams,
elabExeDependencyPaths,
elabLibDependencies,
elabExeDependencies,
elabSetupDependencies,
elabPkgConfigDependencies,
ElaboratedPackageOrComponent(..),
ElaboratedComponent(..),
ElaboratedPackage(..),
pkgOrderDependencies,
ElaboratedPlanPackage,
ElaboratedSharedConfig(..),
ElaboratedReadyPackage,
BuildStyle(..),
CabalFileText,
-- * Build targets
PackageTarget(..),
ComponentTarget(..),
showComponentTarget,
SubComponentTarget(..),
-- * Setup script
SetupScriptStyle(..),
) where
import Distribution.Client.PackageHash
import Distribution.Client.Types
import Distribution.Client.InstallPlan
( GenericInstallPlan, GenericPlanPackage )
import Distribution.Client.SolverInstallPlan
( SolverInstallPlan )
import Distribution.Client.DistDirLayout
import Distribution.Types.ComponentEnabledSpec
import Distribution.Package
hiding (InstalledPackageId, installedPackageId)
import Distribution.System
import qualified Distribution.PackageDescription as Cabal
import Distribution.InstalledPackageInfo (InstalledPackageInfo)
import Distribution.Simple.Compiler
import qualified Distribution.Simple.BuildTarget as Cabal
import Distribution.Simple.Program.Db
import Distribution.ModuleName (ModuleName)
import Distribution.Simple.LocalBuildInfo (ComponentName(..))
import qualified Distribution.Simple.InstallDirs as InstallDirs
import Distribution.Simple.InstallDirs (PathTemplate)
import Distribution.Version
import qualified Distribution.Solver.Types.ComponentDeps as CD
import Distribution.Solver.Types.ComponentDeps (ComponentDeps)
import Distribution.Solver.Types.OptionalStanza
import Distribution.Compat.Graph (IsNode(..))
import Distribution.Simple.Utils (ordNub)
import Data.Map (Map)
import Data.Set (Set)
import qualified Data.ByteString.Lazy as LBS
import Distribution.Compat.Binary
import GHC.Generics (Generic)
import qualified Data.Monoid as Mon
-- | The combination of an elaborated install plan plus a
-- 'ElaboratedSharedConfig' contains all the details necessary to be able
-- to execute the plan without having to make further policy decisions.
--
-- It does not include dynamic elements such as resources (such as http
-- connections).
--
type ElaboratedInstallPlan
= GenericInstallPlan InstalledPackageInfo
ElaboratedConfiguredPackage
type ElaboratedPlanPackage
= GenericPlanPackage InstalledPackageInfo
ElaboratedConfiguredPackage
--TODO: [code cleanup] decide if we really need this, there's not much in it, and in principle
-- even platform and compiler could be different if we're building things
-- like a server + client with ghc + ghcjs
data ElaboratedSharedConfig
= ElaboratedSharedConfig {
pkgConfigPlatform :: Platform,
pkgConfigCompiler :: Compiler, --TODO: [code cleanup] replace with CompilerInfo
-- | The programs that the compiler configured (e.g. for GHC, the progs
-- ghc & ghc-pkg). Once constructed, only the 'configuredPrograms' are
-- used.
pkgConfigCompilerProgs :: ProgramDb
}
deriving (Show, Generic)
--TODO: [code cleanup] no Eq instance
instance Binary ElaboratedSharedConfig
data ElaboratedConfiguredPackage
= ElaboratedConfiguredPackage {
-- | The 'UnitId' which uniquely identifies this item in a build plan
elabUnitId :: UnitId,
-- | The 'PackageId' of the originating package
elabPkgSourceId :: PackageId,
-- | Mapping from 'PackageName's to 'ComponentName', for every
-- package that is overloaded with an internal component name
elabInternalPackages :: Map PackageName ComponentName,
-- | A total flag assignment for the package.
-- TODO: Actually this can be per-component if we drop
-- all flags that don't affect a component.
elabFlagAssignment :: Cabal.FlagAssignment,
-- | The original default flag assignment, used only for reporting.
elabFlagDefaults :: Cabal.FlagAssignment,
elabPkgDescription :: Cabal.PackageDescription,
-- | Where the package comes from, e.g. tarball, local dir etc. This
-- is not the same as where it may be unpacked to for the build.
elabPkgSourceLocation :: PackageLocation (Maybe FilePath),
-- | The hash of the source, e.g. the tarball. We don't have this for
-- local source dir packages.
elabPkgSourceHash :: Maybe PackageSourceHash,
-- | Is this package one of the ones specified by location in the
-- project file? (As opposed to a dependency, or a named package pulled
-- in)
elabLocalToProject :: Bool,
-- | Are we going to build and install this package to the store, or are
-- we going to build it and register it locally.
elabBuildStyle :: BuildStyle,
-- | Another way of phrasing 'pkgStanzasAvailable'.
elabEnabledSpec :: ComponentEnabledSpec,
-- | Which optional stanzas (ie testsuites, benchmarks) can be built.
-- This means the solver produced a plan that has them available.
-- This doesn't necessary mean we build them by default.
elabStanzasAvailable :: Set OptionalStanza,
-- | Which optional stanzas the user explicitly asked to enable or
-- to disable. This tells us which ones we build by default, and
-- helps with error messages when the user asks to build something
-- they explicitly disabled.
--
-- TODO: The 'Bool' here should be refined into an ADT with three
-- cases: NotRequested, ExplicitlyRequested and
-- ImplicitlyRequested. A stanza is explicitly requested if
-- the user asked, for this *specific* package, that the stanza
-- be enabled; it's implicitly requested if the user asked for
-- all global packages to have this stanza enabled. The
-- difference between an explicit and implicit request is
-- error reporting behavior: if a user asks for tests to be
-- enabled for a specific package that doesn't have any tests,
-- we should warn them about it, but we shouldn't complain
-- that a user enabled tests globally, and some local packages
-- just happen not to have any tests. (But perhaps we should
-- warn if ALL local packages don't have any tests.)
elabStanzasRequested :: Map OptionalStanza Bool,
elabSetupPackageDBStack :: PackageDBStack,
elabBuildPackageDBStack :: PackageDBStack,
elabRegisterPackageDBStack :: PackageDBStack,
-- | The package/component contains/is a library and so must be registered
elabRequiresRegistration :: Bool,
elabPkgDescriptionOverride :: Maybe CabalFileText,
-- TODO: make per-component variants of these flags
elabVanillaLib :: Bool,
elabSharedLib :: Bool,
elabDynExe :: Bool,
elabGHCiLib :: Bool,
elabProfLib :: Bool,
elabProfExe :: Bool,
elabProfLibDetail :: ProfDetailLevel,
elabProfExeDetail :: ProfDetailLevel,
elabCoverage :: Bool,
elabOptimization :: OptimisationLevel,
elabSplitObjs :: Bool,
elabStripLibs :: Bool,
elabStripExes :: Bool,
elabDebugInfo :: DebugInfoLevel,
elabProgramPaths :: Map String FilePath,
elabProgramArgs :: Map String [String],
elabProgramPathExtra :: [FilePath],
elabConfigureScriptArgs :: [String],
elabExtraLibDirs :: [FilePath],
elabExtraFrameworkDirs :: [FilePath],
elabExtraIncludeDirs :: [FilePath],
elabProgPrefix :: Maybe PathTemplate,
elabProgSuffix :: Maybe PathTemplate,
elabInstallDirs :: InstallDirs.InstallDirs FilePath,
elabHaddockHoogle :: Bool,
elabHaddockHtml :: Bool,
elabHaddockHtmlLocation :: Maybe String,
elabHaddockExecutables :: Bool,
elabHaddockTestSuites :: Bool,
elabHaddockBenchmarks :: Bool,
elabHaddockInternal :: Bool,
elabHaddockCss :: Maybe FilePath,
elabHaddockHscolour :: Bool,
elabHaddockHscolourCss :: Maybe FilePath,
elabHaddockContents :: Maybe PathTemplate,
-- Setup.hs related things:
-- | One of four modes for how we build and interact with the Setup.hs
-- script, based on whether it's a build-type Custom, with or without
-- explicit deps and the cabal spec version the .cabal file needs.
elabSetupScriptStyle :: SetupScriptStyle,
-- | The version of the Cabal command line interface that we are using
-- for this package. This is typically the version of the Cabal lib
-- that the Setup.hs is built against.
elabSetupScriptCliVersion :: Version,
-- Build time related:
elabBuildTargets :: [ComponentTarget],
elabReplTarget :: Maybe ComponentTarget,
elabBuildHaddocks :: Bool,
--pkgSourceDir ? -- currently passed in later because they can use temp locations
--pkgBuildDir ? -- but could in principle still have it here, with optional instr to use temp loc
-- | Component/package specific information
elabPkgOrComp :: ElaboratedPackageOrComponent
}
deriving (Eq, Show, Generic)
instance Package ElaboratedConfiguredPackage where
packageId = elabPkgSourceId
instance HasConfiguredId ElaboratedConfiguredPackage where
configuredId elab = ConfiguredId (packageId elab) (unitIdComponentId (elabUnitId elab))
instance HasUnitId ElaboratedConfiguredPackage where
installedUnitId = elabUnitId
instance IsNode ElaboratedConfiguredPackage where
type Key ElaboratedConfiguredPackage = UnitId
nodeKey = elabUnitId
nodeNeighbors elab = case elabPkgOrComp elab of
-- Important not to have duplicates: otherwise InstallPlan gets
-- confused. NB: this DOES include setup deps.
ElabPackage pkg -> ordNub (CD.flatDeps (pkgOrderDependencies pkg))
ElabComponent comp -> compOrderDependencies comp
instance Binary ElaboratedConfiguredPackage
data ElaboratedPackageOrComponent
= ElabPackage ElaboratedPackage
| ElabComponent ElaboratedComponent
deriving (Eq, Show, Generic)
instance Binary ElaboratedPackageOrComponent
elabDistDirParams :: ElaboratedSharedConfig -> ElaboratedConfiguredPackage -> DistDirParams
elabDistDirParams shared elab = DistDirParams {
distParamUnitId = installedUnitId elab,
distParamPackageId = elabPkgSourceId elab,
distParamComponentName = case elabPkgOrComp elab of
ElabComponent comp -> compComponentName comp
ElabPackage _ -> Nothing,
distParamCompilerId = compilerId (pkgConfigCompiler shared),
distParamPlatform = pkgConfigPlatform shared,
distParamOptimization = elabOptimization elab
}
-- | The library dependencies (i.e., the libraries we depend on, NOT
-- the dependencies of the library), NOT including setup dependencies.
elabLibDependencies :: ElaboratedConfiguredPackage -> [ConfiguredId]
elabLibDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabPackage pkg }
= ordNub (CD.nonSetupDeps (pkgLibDependencies pkg))
elabLibDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabComponent comp }
= compLibDependencies comp
elabExeDependencies :: ElaboratedConfiguredPackage -> [ComponentId]
elabExeDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabPackage pkg }
= map confInstId (CD.nonSetupDeps (pkgExeDependencies pkg))
elabExeDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabComponent comp }
= compExeDependencies comp
elabExeDependencyPaths :: ElaboratedConfiguredPackage -> [FilePath]
elabExeDependencyPaths ElaboratedConfiguredPackage { elabPkgOrComp = ElabPackage pkg }
= CD.nonSetupDeps (pkgExeDependencyPaths pkg)
elabExeDependencyPaths ElaboratedConfiguredPackage { elabPkgOrComp = ElabComponent comp }
= compExeDependencyPaths comp
elabSetupDependencies :: ElaboratedConfiguredPackage -> [ConfiguredId]
elabSetupDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabPackage pkg }
= CD.setupDeps (pkgLibDependencies pkg)
elabSetupDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabComponent comp }
= compSetupDependencies comp
elabPkgConfigDependencies :: ElaboratedConfiguredPackage -> [(PackageName, Maybe Version)]
elabPkgConfigDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabPackage pkg }
= pkgPkgConfigDependencies pkg
elabPkgConfigDependencies ElaboratedConfiguredPackage { elabPkgOrComp = ElabComponent comp }
= compPkgConfigDependencies comp
-- | Some extra metadata associated with an
-- 'ElaboratedConfiguredPackage' which indicates that the "package"
-- in question is actually a single component to be built. Arguably
-- it would be clearer if there were an ADT which branched into
-- package work items and component work items, but I've structured
-- it this way to minimize change to the existing code (which I
-- don't feel qualified to rewrite.)
data ElaboratedComponent
= ElaboratedComponent {
-- | The name of the component to be built according to the solver
compSolverName :: CD.Component,
-- | The name of the component to be built. Nothing if
-- it's a setup dep.
compComponentName :: Maybe ComponentName,
-- | The library dependencies of this component.
compLibDependencies :: [ConfiguredId],
-- | The executable dependencies of this component.
compExeDependencies :: [ComponentId],
-- | The @pkg-config@ dependencies of the component
compPkgConfigDependencies :: [(PackageName, Maybe Version)],
-- | The paths all our executable dependencies will be installed
-- to once they are installed.
compExeDependencyPaths :: [FilePath],
-- | The setup dependencies. TODO: Remove this when setups
-- are components of their own.
compSetupDependencies :: [ConfiguredId]
}
deriving (Eq, Show, Generic)
instance Binary ElaboratedComponent
compOrderDependencies :: ElaboratedComponent -> [UnitId]
compOrderDependencies comp =
-- TODO: Change this with Backpack!
map (SimpleUnitId . confInstId) (compLibDependencies comp)
++ map SimpleUnitId (compExeDependencies comp)
++ map (SimpleUnitId . confInstId) (compSetupDependencies comp)
data ElaboratedPackage
= ElaboratedPackage {
pkgInstalledId :: InstalledPackageId,
-- | The exact dependencies (on other plan packages)
--
pkgLibDependencies :: ComponentDeps [ConfiguredId],
-- | Dependencies on executable packages.
--
pkgExeDependencies :: ComponentDeps [ConfiguredId],
-- | Paths where executable dependencies live.
--
pkgExeDependencyPaths :: ComponentDeps [FilePath],
-- | Dependencies on @pkg-config@ packages.
-- NB: this is NOT per-component (although it could be)
-- because Cabal library does not track per-component
-- pkg-config depends; it always does them all at once.
--
pkgPkgConfigDependencies :: [(PackageName, Maybe Version)],
-- | Which optional stanzas (ie testsuites, benchmarks) will actually
-- be enabled during the package configure step.
pkgStanzasEnabled :: Set OptionalStanza
}
deriving (Eq, Show, Generic)
instance Binary ElaboratedPackage
pkgOrderDependencies :: ElaboratedPackage -> ComponentDeps [UnitId]
pkgOrderDependencies pkg =
fmap (map (SimpleUnitId . confInstId)) (pkgLibDependencies pkg) `Mon.mappend`
fmap (map (SimpleUnitId . confInstId)) (pkgExeDependencies pkg)
-- | This is used in the install plan to indicate how the package will be
-- built.
--
data BuildStyle =
-- | The classic approach where the package is built, then the files
-- installed into some location and the result registered in a package db.
--
-- If the package came from a tarball then it's built in a temp dir and
-- the results discarded.
BuildAndInstall
-- | The package is built, but the files are not installed anywhere,
-- rather the build dir is kept and the package is registered inplace.
--
-- Such packages can still subsequently be installed.
--
-- Typically 'BuildAndInstall' packages will only depend on other
-- 'BuildAndInstall' style packages and not on 'BuildInplaceOnly' ones.
--
| BuildInplaceOnly
deriving (Eq, Show, Generic)
instance Binary BuildStyle
type CabalFileText = LBS.ByteString
type ElaboratedReadyPackage = GenericReadyPackage ElaboratedConfiguredPackage
---------------------------
-- Build targets
--
-- | The various targets within a package. This is more of a high level
-- specification than a elaborated prescription.
--
data PackageTarget =
-- | Build the default components in this package. This usually means
-- just the lib and exes, but it can also mean the testsuites and
-- benchmarks if the user explicitly requested them.
BuildDefaultComponents
-- | Build a specific component in this package.
| BuildSpecificComponent ComponentTarget
| ReplDefaultComponent
| ReplSpecificComponent ComponentTarget
| HaddockDefaultComponents
deriving (Eq, Show, Generic)
data ComponentTarget = ComponentTarget ComponentName SubComponentTarget
deriving (Eq, Ord, Show, Generic)
data SubComponentTarget = WholeComponent
| ModuleTarget ModuleName
| FileTarget FilePath
deriving (Eq, Ord, Show, Generic)
instance Binary PackageTarget
instance Binary ComponentTarget
instance Binary SubComponentTarget
-- | Unambiguously render a 'ComponentTarget', e.g., to pass
-- to a Cabal Setup script.
showComponentTarget :: PackageId -> ComponentTarget -> String
showComponentTarget pkgid =
Cabal.showBuildTarget pkgid . toBuildTarget
where
toBuildTarget :: ComponentTarget -> Cabal.BuildTarget
toBuildTarget (ComponentTarget cname subtarget) =
case subtarget of
WholeComponent -> Cabal.BuildTargetComponent cname
ModuleTarget mname -> Cabal.BuildTargetModule cname mname
FileTarget fname -> Cabal.BuildTargetFile cname fname
---------------------------
-- Setup.hs script policy
--
-- | There are four major cases for Setup.hs handling:
--
-- 1. @build-type@ Custom with a @custom-setup@ section
-- 2. @build-type@ Custom without a @custom-setup@ section
-- 3. @build-type@ not Custom with @cabal-version > $our-cabal-version@
-- 4. @build-type@ not Custom with @cabal-version <= $our-cabal-version@
--
-- It's also worth noting that packages specifying @cabal-version: >= 1.23@
-- or later that have @build-type@ Custom will always have a @custom-setup@
-- section. Therefore in case 2, the specified @cabal-version@ will always be
-- less than 1.23.
--
-- In cases 1 and 2 we obviously have to build an external Setup.hs script,
-- while in case 4 we can use the internal library API. In case 3 we also have
-- to build an external Setup.hs script because the package needs a later
-- Cabal lib version than we can support internally.
--
data SetupScriptStyle = SetupCustomExplicitDeps
| SetupCustomImplicitDeps
| SetupNonCustomExternalLib
| SetupNonCustomInternalLib
deriving (Eq, Show, Generic)
instance Binary SetupScriptStyle
| sopvop/cabal | cabal-install/Distribution/Client/ProjectPlanning/Types.hs | bsd-3-clause | 20,460 | 0 | 13 | 4,730 | 2,394 | 1,434 | 960 | 255 | 3 |
-- Copyright (c) 2016-present, Facebook, Inc.
-- All rights reserved.
--
-- This source code is licensed under the BSD-style license found in the
-- LICENSE file in the root directory of this source tree. An additional grant
-- of patent rights can be found in the PATENTS file in the same directory.
{-# LANGUAGE OverloadedStrings #-}
module Duckling.Time.NB.Corpus
( corpus ) where
import Prelude
import Data.String
import Duckling.Lang
import Duckling.Resolve
import Duckling.Time.Corpus
import Duckling.Time.Types hiding (Month)
import Duckling.TimeGrain.Types hiding (add)
import Duckling.Testing.Types hiding (examples)
corpus :: Corpus
corpus = (testContext {lang = NB}, allExamples)
allExamples :: [Example]
allExamples = concat
[ examples (datetime (2013, 2, 12, 4, 30, 0) Second)
[ "nå"
, "akkurat nå"
]
, examples (datetime (2013, 2, 12, 0, 0, 0) Day)
[ "i dag"
]
, examples (datetime (2013, 2, 11, 0, 0, 0) Day)
[ "i går"
]
, examples (datetime (2013, 2, 13, 0, 0, 0) Day)
[ "i morgen"
]
, examples (datetime (2013, 2, 18, 0, 0, 0) Day)
[ "mandag"
, "man."
, "på mandag"
]
, examples (datetime (2013, 2, 18, 0, 0, 0) Day)
[ "Mandag den 18. februar"
, "Man, 18 februar"
]
, examples (datetime (2013, 2, 19, 0, 0, 0) Day)
[ "tirsdag"
]
, examples (datetime (2013, 2, 14, 0, 0, 0) Day)
[ "torsdag"
, "tors"
, "tors."
]
, examples (datetime (2013, 2, 15, 0, 0, 0) Day)
[ "fredag"
, "fre"
, "fre."
]
, examples (datetime (2013, 2, 16, 0, 0, 0) Day)
[ "lørdag"
, "lør"
, "lør."
]
, examples (datetime (2013, 2, 17, 0, 0, 0) Day)
[ "søndag"
, "søn"
, "søn."
]
, examples (datetime (2013, 3, 1, 0, 0, 0) Day)
[ "Den første mars"
, "1. mars"
, "Den 1. mars"
]
, examples (datetime (2013, 3, 3, 0, 0, 0) Day)
[ "3 mars"
, "den tredje mars"
, "den 3. mars"
]
, examples (datetime (2015, 3, 3, 0, 0, 0) Day)
[ "3 mars 2015"
, "tredje mars 2015"
, "3. mars 2015"
, "3-3-2015"
, "03-03-2015"
, "3/3/2015"
, "3/3/15"
, "2015-3-3"
, "2015-03-03"
]
, examples (datetime (2013, 2, 15, 0, 0, 0) Day)
[ "På den 15."
, "På den 15"
, "Den 15."
, "Den femtende"
]
, examples (datetime (2013, 2, 15, 0, 0, 0) Day)
[ "den 15. februar"
, "15. februar"
, "februar 15"
, "15-02"
, "15/02"
]
, examples (datetime (2013, 8, 8, 0, 0, 0) Day)
[ "8 Aug"
]
, examples (datetime (2014, 10, 0, 0, 0, 0) Month)
[ "Oktober 2014"
]
, examples (datetime (1974, 10, 31, 0, 0, 0) Day)
[ "31/10/1974"
, "31/10/74"
, "31-10-74"
]
, examples (datetime (2015, 4, 14, 0, 0, 0) Day)
[ "14april 2015"
, "April 14, 2015"
, "fjortende April 15"
]
, examples (datetime (2013, 2, 22, 0, 0, 0) Day)
[ "neste fredag igjen"
]
, examples (datetime (2013, 3, 0, 0, 0, 0) Month)
[ "neste mars"
]
, examples (datetime (2014, 3, 0, 0, 0, 0) Month)
[ "neste mars igjen"
]
, examples (datetime (2013, 2, 10, 0, 0, 0) Day)
[ "Søndag, 10 feb"
, "Søndag 10 Feb"
]
, examples (datetime (2013, 2, 13, 0, 0, 0) Day)
[ "Ons, Feb13"
, "Ons feb13"
]
, examples (datetime (2013, 2, 18, 0, 0, 0) Day)
[ "Mandag, Feb 18"
, "Man, februar 18"
]
, examples (datetime (2013, 2, 11, 0, 0, 0) Week)
[ "denne uken"
]
, examples (datetime (2013, 2, 4, 0, 0, 0) Week)
[ "forrige uke"
, "sist uke"
]
, examples (datetime (2013, 2, 18, 0, 0, 0) Week)
[ "neste uke"
]
, examples (datetime (2013, 1, 0, 0, 0, 0) Month)
[ "forrige måned"
, "sist måned"
]
, examples (datetime (2013, 3, 0, 0, 0, 0) Month)
[ "neste måned"
]
, examples (datetime (2013, 1, 1, 0, 0, 0) Quarter)
[ "dette kvartalet"
]
, examples (datetime (2013, 4, 1, 0, 0, 0) Quarter)
[ "neste kvartal"
]
, examples (datetime (2013, 7, 1, 0, 0, 0) Quarter)
[ "tredje kvartal"
, "3. kvartal"
]
, examples (datetime (2018, 10, 1, 0, 0, 0) Quarter)
[ "4. kvartal 2018"
, "fjerde kvartal 2018"
]
, examples (datetime (2012, 0, 0, 0, 0, 0) Year)
[ "forrige år"
, "sist år"
]
, examples (datetime (2012, 0, 0, 0, 0, 0) Year)
[ "i fjor"
]
, examples (datetime (2013, 0, 0, 0, 0, 0) Year)
[ "i år"
, "dette år"
]
, examples (datetime (2014, 0, 0, 0, 0, 0) Year)
[ "neste år"
]
, examples (datetime (2013, 2, 10, 0, 0, 0) Day)
[ "forrige søndag"
, "sist søndag"
, "søndag i forrige uke"
]
, examples (datetime (2013, 2, 5, 0, 0, 0) Day)
[ "forrige tirsdag"
, "sist tirsdag"
]
, examples (datetime (2013, 2, 19, 0, 0, 0) Day)
[ "neste tirsdag"
]
, examples (datetime (2013, 2, 13, 0, 0, 0) Day)
[ "neste onsdag"
]
, examples (datetime (2013, 2, 20, 0, 0, 0) Day)
[ "onsdag i neste uke"
, "onsdag neste uke"
, "neste onsdag igjen"
]
, examples (datetime (2013, 2, 22, 0, 0, 0) Day)
[ "neste fredag igjen"
]
, examples (datetime (2013, 2, 11, 0, 0, 0) Day)
[ "mandag denne uken"
]
, examples (datetime (2013, 2, 12, 0, 0, 0) Day)
[ "tirsdag denne uken"
]
, examples (datetime (2013, 2, 13, 0, 0, 0) Day)
[ "onsdag denne uken"
]
, examples (datetime (2013, 2, 14, 0, 0, 0) Day)
[ "i overimorgen"
]
, examples (datetime (2013, 2, 10, 0, 0, 0) Day)
[ "i forigårs"
]
, examples (datetime (2013, 3, 25, 0, 0, 0) Day)
[ "siste mandag i mars"
, "siste mandag i mars"
]
, examples (datetime (2014, 3, 30, 0, 0, 0) Day)
[ "siste søndag i mars 2014"
, "siste søndag i mars 2014"
]
, examples (datetime (2013, 10, 3, 0, 0, 0) Day)
[ "tredje dag i oktober"
, "tredje dag i Oktober"
]
, examples (datetime (2014, 10, 6, 0, 0, 0) Week)
[ "første uke i oktober 2014"
, "første uke i Oktober 2014"
]
, examples (datetime (2015, 10, 31, 0, 0, 0) Day)
[ "siste dag i oktober 2015"
, "siste dag i Oktober 2015"
]
, examples (datetime (2014, 9, 22, 0, 0, 0) Week)
[ "siste uke i september 2014"
, "siste uke i September 2014"
]
, examples (datetime (2013, 10, 1, 0, 0, 0) Day)
[ "første tirsdag i oktober"
, "første tirsdag i Oktober"
]
, examples (datetime (2014, 9, 16, 0, 0, 0) Day)
[ "tredje tirsdag i september 2014"
, "tredje tirsdag i September 2014"
]
, examples (datetime (2014, 10, 1, 0, 0, 0) Day)
[ "første onsdag i oktober 2014"
, "første onsdag i Oktober 2014"
]
, examples (datetime (2014, 10, 8, 0, 0, 0) Day)
[ "andre onsdag i oktober 2014"
, "andre onsdag i Oktober 2014"
]
, examples (datetime (2013, 2, 13, 3, 0, 0) Hour)
[ "klokken 3"
, "kl. 3"
]
, examples (datetime (2013, 2, 13, 3, 18, 0) Minute)
[ "3:18"
]
, examples (datetime (2013, 2, 12, 15, 0, 0) Hour)
[ "klokken 15"
, "kl. 15"
, "15h"
]
, examples (datetime (2013, 2, 12, 15, 0, 0) Hour)
[ "ca. kl. 15"
, "cirka kl. 15"
, "omkring klokken 15"
]
, examples (datetime (2013, 2, 13, 17, 0, 0) Hour)
[ "imorgen klokken 17 sharp"
, "imorgen kl. 17 presis"
]
, examples (datetime (2013, 2, 12, 15, 15, 0) Minute)
[ "kvarter over 15"
, "kvart over 15"
, "15:15"
]
, examples (datetime (2013, 2, 12, 15, 20, 0) Minute)
[ "kl. 20 over 15"
, "klokken 20 over 15"
, "kl. 15:20"
, "15:20"
]
, examples (datetime (2013, 2, 12, 15, 30, 0) Minute)
[ "15:30"
]
, examples (datetime (2013, 2, 12, 15, 23, 24) Second)
[ "15:23:24"
]
, examples (datetime (2013, 2, 12, 11, 45, 0) Minute)
[ "kvarter på 12"
, "kvart på 12"
, "11:45"
]
, examples (datetime (2013, 2, 16, 9, 0, 0) Hour)
[ "klokken 9 på lørdag"
]
, examples (datetime (2014, 7, 18, 19, 0, 0) Minute)
[ "Fre, Jul 18, 2014 19:00"
]
, examples (datetime (2014, 7, 18, 0, 0, 0) Day)
[ "Fre, Jul 18"
, "Jul 18, Fre"
]
, examples (datetime (2014, 9, 20, 19, 30, 0) Minute)
[ "kl. 19:30, Lør, 20 sep"
]
, examples (datetime (2013, 2, 12, 4, 30, 1) Second)
[ "om 1 sekund"
, "om ett sekund"
, "om et sekund"
, "ett sekund fra nå"
, "et sekund fra nå"
]
, examples (datetime (2013, 2, 12, 4, 31, 0) Second)
[ "om 1 minutt"
, "om et minutt"
, "om ett minutt"
]
, examples (datetime (2013, 2, 12, 4, 32, 0) Second)
[ "om 2 minutter"
, "om to minutter"
, "om 2 minutter mer"
, "om to minutter mer"
, "2 minutter fra nå"
, "to minutter fra nå"
]
, examples (datetime (2013, 2, 12, 5, 30, 0) Second)
[ "om 60 minutter"
]
, examples (datetime (2013, 2, 12, 5, 0, 0) Second)
[ "om en halv time"
]
, examples (datetime (2013, 2, 12, 7, 0, 0) Second)
[ "om 2,5 time"
, "om 2 og en halv time"
, "om to og en halv time"
]
, examples (datetime (2013, 2, 12, 5, 30, 0) Minute)
[ "om én time"
, "om 1 time"
, "om 1t"
]
, examples (datetime (2013, 2, 12, 6, 30, 0) Minute)
[ "om et par timer"
]
, examples (datetime (2013, 2, 13, 4, 30, 0) Minute)
[ "om 24 timer"
]
, examples (datetime (2013, 2, 13, 4, 0, 0) Hour)
[ "om en dag"
]
, examples (datetime (2016, 2, 0, 0, 0, 0) Month)
[ "3 år fra i dag"
]
, examples (datetime (2013, 2, 19, 4, 0, 0) Hour)
[ "om 7 dager"
]
, examples (datetime (2013, 2, 19, 0, 0, 0) Day)
[ "om en uke"
, "om én uke"
]
, examples (datetime (2013, 2, 12, 5, 0, 0) Second)
[ "om ca. en halv time"
, "om cirka en halv time"
]
, examples (datetime (2013, 2, 5, 4, 0, 0) Hour)
[ "7 dager siden"
, "syv dager siden"
]
, examples (datetime (2013, 1, 29, 4, 0, 0) Hour)
[ "14 dager siden"
, "fjorten dager siden"
]
, examples (datetime (2013, 2, 5, 0, 0, 0) Day)
[ "en uke siden"
, "én uke siden"
, "1 uke siden"
]
, examples (datetime (2013, 1, 22, 0, 0, 0) Day)
[ "3 uker siden"
, "tre uker siden"
]
, examples (datetime (2012, 11, 12, 0, 0, 0) Day)
[ "3 måneder siden"
, "tre måneder siden"
]
, examples (datetime (2011, 2, 0, 0, 0, 0) Month)
[ "to år siden"
, "2 år siden"
]
, examples (datetime (1954, 0, 0, 0, 0, 0) Year)
[ "1954"
]
, examples (datetime (2013, 12, 0, 0, 0, 0) Month)
[ "et år etter julaften"
, "ett år etter julaften"
]
, examples (datetimeInterval ((2013, 6, 21, 0, 0, 0), (2013, 9, 24, 0, 0, 0)) Day)
[ "denne sommeren"
, "den her sommeren"
]
, examples (datetimeInterval ((2012, 12, 21, 0, 0, 0), (2013, 3, 21, 0, 0, 0)) Day)
[ "denne vinteren"
, "den her vinteren"
]
, examples (datetime (2013, 12, 25, 0, 0, 0) Day)
[ "1 juledag"
, "1. juledag"
, "første juledag"
]
, examples (datetime (2013, 12, 31, 0, 0, 0) Day)
[ "nyttårsaften"
]
, examples (datetime (2014, 1, 1, 0, 0, 0) Day)
[ "nyttårsdag"
]
, examples (datetimeInterval ((2013, 2, 12, 18, 0, 0), (2013, 2, 13, 0, 0, 0)) Hour)
[ "i kveld"
]
, examples (datetimeInterval ((2013, 2, 8, 18, 0, 0), (2013, 2, 11, 0, 0, 0)) Hour)
[ "forrige helg"
, "sist helg"
]
, examples (datetimeInterval ((2013, 2, 13, 18, 0, 0), (2013, 2, 14, 0, 0, 0)) Hour)
[ "i morgen kveld"
]
, examples (datetimeInterval ((2013, 2, 13, 12, 0, 0), (2013, 2, 13, 14, 0, 0)) Hour)
[ "i morgen middag"
]
, examples (datetimeInterval ((2013, 2, 11, 18, 0, 0), (2013, 2, 12, 0, 0, 0)) Hour)
[ "i går kveld"
]
, examples (datetimeInterval ((2013, 2, 15, 18, 0, 0), (2013, 2, 18, 0, 0, 0)) Hour)
[ "denne helgen"
, "denne helga"
, "i helga"
, "i helgen"
]
, examples (datetimeInterval ((2013, 2, 18, 4, 0, 0), (2013, 2, 18, 12, 0, 0)) Hour)
[ "mandag morgen"
]
, examples (datetimeInterval ((2013, 12, 24, 0, 0, 0), (2013, 12, 31, 0, 0, 0)) Day)
[ "i romjulen"
, "i romjula"
]
, examples (datetimeInterval ((2013, 2, 12, 4, 29, 58), (2013, 2, 12, 4, 30, 0)) Second)
[ "siste 2 sekunder"
, "siste to sekunder"
]
, examples (datetimeInterval ((2013, 2, 12, 4, 30, 1), (2013, 2, 12, 4, 30, 4)) Second)
[ "neste 3 sekunder"
, "neste tre sekunder"
]
, examples (datetimeInterval ((2013, 2, 12, 4, 28, 0), (2013, 2, 12, 4, 30, 0)) Minute)
[ "siste 2 minutter"
, "siste to minutter"
]
, examples (datetimeInterval ((2013, 2, 12, 4, 31, 0), (2013, 2, 12, 4, 34, 0)) Minute)
[ "neste 3 minutter"
, "neste tre minutter"
]
, examples (datetimeInterval ((2013, 2, 12, 3, 0, 0), (2013, 2, 12, 4, 0, 0)) Hour)
[ "siste 1 time"
, "seneste 1 time"
]
, examples (datetimeInterval ((2013, 2, 12, 5, 0, 0), (2013, 2, 12, 8, 0, 0)) Hour)
[ "neste 3 timer"
, "neste tre timer"
]
, examples (datetimeInterval ((2013, 2, 10, 0, 0, 0), (2013, 2, 12, 0, 0, 0)) Day)
[ "siste 2 dager"
, "siste to dager"
, "seneste 2 dager"
]
, examples (datetimeInterval ((2013, 2, 13, 0, 0, 0), (2013, 2, 16, 0, 0, 0)) Day)
[ "neste 3 dager"
, "neste tre dager"
]
, examples (datetimeInterval ((2013, 1, 28, 0, 0, 0), (2013, 2, 11, 0, 0, 0)) Week)
[ "siste 2 uker"
, "siste to uker"
, "seneste to uker"
]
, examples (datetimeInterval ((2013, 2, 18, 0, 0, 0), (2013, 3, 11, 0, 0, 0)) Week)
[ "neste 3 uker"
, "neste tre uker"
]
, examples (datetimeInterval ((2012, 12, 0, 0, 0, 0), (2013, 2, 0, 0, 0, 0)) Month)
[ "siste 2 måneder"
, "siste to måneder"
, "seneste to måneder"
]
, examples (datetimeInterval ((2013, 3, 0, 0, 0, 0), (2013, 6, 0, 0, 0, 0)) Month)
[ "neste 3 måneder"
, "neste tre måneder"
]
, examples (datetimeInterval ((2011, 0, 0, 0, 0, 0), (2013, 0, 0, 0, 0, 0)) Year)
[ "siste 2 år"
, "siste to år"
, "seneste 2 år"
]
, examples (datetimeInterval ((2014, 0, 0, 0, 0, 0), (2017, 0, 0, 0, 0, 0)) Year)
[ "neste 3 år"
, "neste tre år"
]
, examples (datetimeInterval ((2013, 7, 13, 0, 0, 0), (2013, 7, 16, 0, 0, 0)) Day)
[ "13-15 juli"
, "13-15 Juli"
, "13 til 15 Juli"
, "13 juli til 15 juli"
]
, examples (datetimeInterval ((2013, 8, 8, 0, 0, 0), (2013, 8, 13, 0, 0, 0)) Day)
[ "8 Aug - 12 Aug"
, "8 Aug - 12 aug"
, "8 aug - 12 aug"
, "8 august - 12 august"
]
, examples (datetimeInterval ((2013, 2, 12, 9, 30, 0), (2013, 2, 12, 11, 1, 0)) Minute)
[ "9:30 - 11:00"
, "9:30 til 11:00"
]
, examples (datetimeInterval ((2013, 2, 14, 9, 30, 0), (2013, 2, 14, 11, 1, 0)) Minute)
[ "fra 9:30 - 11:00 på torsdag"
, "fra 9:30 til 11:00 på torsdag"
, "mellom 9:30 og 11:00 på torsdag"
, "9:30 - 11:00 på torsdag"
, "9:30 til 11:00 på torsdag"
, "etter 9:30 men før 11:00 på torsdag"
, "torsdag fra 9:30 til 11:00"
, "torsdag mellom 9:30 og 11:00"
, "fra 9:30 til 11:00 på torsdag"
]
, examples (datetimeInterval ((2013, 2, 14, 9, 0, 0), (2013, 2, 14, 12, 0, 0)) Hour)
[ "torsdag fra 9 til 11"
]
, examples (datetimeInterval ((2013, 2, 12, 11, 30, 0), (2013, 2, 12, 13, 31, 0)) Minute)
[ "11:30-13:30"
]
, examples (datetimeInterval ((2013, 2, 12, 4, 30, 0), (2013, 2, 26, 0, 0, 0)) Second)
[ "innenfor 2 uker"
]
, examples (datetimeOpenInterval Before (2013, 2, 12, 14, 0, 0) Hour)
[ "innen kl. 14"
, "innen klokken 14"
]
, examples (datetime (2013, 2, 12, 13, 0, 0) Minute)
[ "16h CET"
, "kl. 16 CET"
, "klokken 16 CET"
]
, examples (datetime (2013, 2, 14, 6, 0, 0) Minute)
[ "torsdag kl. 8:00 GMT"
, "torsdag klokken 8:00 GMT"
, "torsdag 08:00 GMT"
]
, examples (datetime (2013, 2, 12, 14, 0, 0) Hour)
[ "idag kl. 14"
, "idag klokken 14"
, "kl. 14"
, "klokken 14"
]
, examples (datetime (2013, 4, 25, 16, 0, 0) Minute)
[ "25/4 kl. 16:00"
, "25/4 klokken 16:00"
, "25-04 klokken 16:00"
, "25-4 kl. 16:00"
]
, examples (datetime (2013, 2, 13, 15, 0, 0) Minute)
[ "15:00 i morgen"
, "kl. 15:00 i morgen"
, "klokken 15:00 i morgen"
]
, examples (datetimeOpenInterval After (2013, 2, 12, 14, 0, 0) Hour)
[ "etter kl. 14"
, "etter klokken 14"
]
, examples (datetimeOpenInterval After (2013, 2, 17, 4, 0, 0) Hour)
[ "etter 5 dager"
, "etter fem dager"
]
, examples (datetime (2013, 2, 17, 4, 0, 0) Hour)
[ "om 5 dager"
, "om fem dager"
]
, examples (datetimeOpenInterval After (2013, 2, 13, 14, 0, 0) Hour)
[ "etter i morgen kl. 14"
, "etter i morgen klokken 14"
, "i morgen etter kl. 14"
, "i morgen etter klokken 14"
]
, examples (datetimeOpenInterval Before (2013, 2, 12, 11, 0, 0) Hour)
[ "før kl. 11"
, "før klokken 11"
]
, examples (datetimeOpenInterval Before (2013, 2, 13, 11, 0, 0) Hour)
[ "i morgen før kl. 11"
, "i morgen før klokken 11"
]
, examples (datetimeInterval ((2013, 2, 12, 12, 0, 0), (2013, 2, 12, 19, 0, 0)) Hour)
[ "om ettermiddagen"
]
, examples (datetime (2013, 2, 12, 13, 30, 0) Minute)
[ "kl. 13:30"
, "klokken 13:30"
]
, examples (datetime (2013, 2, 12, 4, 45, 0) Second)
[ "om 15 minutter"
]
, examples (datetimeInterval ((2013, 2, 12, 13, 0, 0), (2013, 2, 12, 17, 0, 0)) Hour)
[ "etter frokost"
]
, examples (datetime (2013, 2, 12, 10, 30, 0) Minute)
[ "10:30"
]
, examples (datetimeInterval ((2013, 2, 12, 4, 0, 0), (2013, 2, 12, 12, 0, 0)) Hour)
[ "denne morgen"
]
, examples (datetime (2013, 2, 18, 0, 0, 0) Day)
[ "neste mandag"
]
, examples (datetime (2014, 2, 9, 0, 0, 0) Day)
[ "morsdag"
]
, examples (datetime (2013, 11, 10, 0, 0, 0) Day)
[ "farsdag"
]
]
| rfranek/duckling | Duckling/Time/NB/Corpus.hs | bsd-3-clause | 22,209 | 0 | 11 | 9,535 | 7,021 | 4,275 | 2,746 | 479 | 1 |
module Test.Property where
import Prelude hiding ((>>))
import Elm.Utils ((|>), (>>))
import Test.Tasty
import Test.Tasty.HUnit
import Test.Tasty.QuickCheck
import Test.QuickCheck.IO ()
import Reporting.Annotation (stripRegion)
import qualified AST.Module
import qualified Data.Text as Text
import qualified Data.Maybe as Maybe
import qualified ElmFormat.Parse as Parse
import qualified ElmFormat.Render.Text as Render
import qualified ElmVersion
import qualified Test.Generators ()
import qualified Test.ElmSourceGenerators
assertStringToString :: String -> Assertion
assertStringToString source =
let
source' = Text.pack source
result =
Parse.parse source'
|> Parse.toEither
|> fmap (Render.render ElmVersion.Elm_0_17)
in
assertEqual "" (Right source') result
astToAst :: AST.Module.Module -> Assertion
astToAst ast =
let
result =
ast
|> Render.render ElmVersion.Elm_0_17
|> Parse.parse
|> Parse.toEither
in
assertEqual ""
(Right $ stripRegion ast)
(fmap stripRegion result)
simpleAst =
case Parse.toEither $ Parse.parse $ Text.pack "module Main exposing (..)\n\n\nfoo =\n 8\n" of
Right ast -> ast
reportFailedAst ast =
let
rendering = Render.render ElmVersion.Elm_0_17 ast |> Text.unpack
result =
Render.render ElmVersion.Elm_0_17 ast
|> Parse.parse
|> fmap stripRegion
|> show
in
concat
[ "=== Parsed as:\n"
, result
, "=== END OF parse\n"
, "=== Rendering of failed AST:\n"
, rendering
, "=== END OF failed AST rendering\n"
]
withCounterexample fn prop =
(\s -> counterexample (fn s) $ prop s)
propertyTests :: TestTree
propertyTests =
testGroup "example test group"
[ testCase "simple AST round trip" $
astToAst simpleAst
, testProperty "rendered AST should parse as equivalent AST"
$ withCounterexample reportFailedAst astToAst
, testGroup "valid Elm files"
[ testProperty "should parse"
$ forAll Test.ElmSourceGenerators.elmModule $ withCounterexample id
$ Text.pack >> Parse.parse >> Parse.toMaybe >> Maybe.isJust
, testProperty "should parse to the same AST after formatting"
$ forAll Test.ElmSourceGenerators.elmModule $ withCounterexample id
$ Text.pack >> Parse.parse >> Parse.toMaybe
>> fmap astToAst
>> Maybe.fromMaybe (assertFailure "failed to parse original")
]
, testCase "simple round trip" $
assertStringToString "module Main exposing (..)\n\n\nfoo =\n 8\n"
, testCase "simple round trip with comments" $
assertStringToString "module Main exposing (..)\n\n\nfoo =\n ( {- A -} 3 {- B -}, {- C -} 4 {- D -} )\n"
, testCase "simple round trip with comments" $
assertStringToString "module Main exposing (..)\n\n\ncommentedLiterals =\n ( {- int -} 1, {- float -} 0.1, {- char -} \'c\', {- string -} \"str\", {- boolean -} True )\n"
]
| nukisman/elm-format-short | tests/Test/Property.hs | bsd-3-clause | 3,229 | 0 | 17 | 943 | 641 | 340 | 301 | 78 | 1 |
{-# LANGUAGE PatternGuards #-}
module Main where
import Tests.Testframe
import System.Environment (getArgs)
import Data.Char (toLower)
import System.Console.GetOpt
import System.Exit (ExitCode(..), exitWith)
import System.IO (hPutStrLn, stderr)
data Flag = Alg String | TestVar String
main = do
args <- getArgs
let argLength = length args
if args !! 1 == "e"
then runExtreme (read $ args !! 2)
else do
let alg = map toLower $ args !! 0
testCase = map toLower $ args !! 1
byteAligned = read $ args !! 2
case lookup testCase testOptions of
Just filePath -> run alg filePath byteAligned
Nothing -> putStrLn $ testCase ++ "is not a valid test file option. \n\nOptions: shortXXX, longXXX"
algorithms = ["jh","groestl"]
testOptions :: [(String,FilePath)]
testOptions = [("short224", "ShortMsgKAT_224.txt"),("short256","ShortMsgKAT_256.txt"),
("short384","ShortMsgKAT_384.txt"),("short512","ShortMsgKAT_512.txt"),
("long224","LongMsgKAT_224.txt"),("long256","LongMsgKAT_256.txt"),
("long384","LongMsgKAT_384.txt"),("long512","LongMsgKAT_512.txt")]
| hakoja/SHA3 | main.hs | bsd-3-clause | 1,174 | 0 | 15 | 257 | 330 | 188 | 142 | 27 | 3 |
module Main where
import Conifer.Types
import Control.Monad (when)
import System.Exit (exitFailure)
import Test.HUnit
main = do
Counts c t e f <- runTestTT tests
when (e > 0 || f > 0) exitFailure
tests = TestList [
TestLabel "getUserDataFromJSON_None" $ TestCase getUserDataFromJSON_None
, TestLabel "getUserDataFromJSON_Age" $ TestCase getUserDataFromJSON_Age
, TestLabel "getUserDataFromJSON_All" $ TestCase getUserDataFromJSON_All
, TestLabel "argsFromInput_NoChange" $ TestCase argsFromInput_NoChange
, TestLabel "argsFromInput_ChangeAll" $ TestCase argsFromInput_ChangeAll
]
getUserDataFromJSON_None = actual @?= expected
where actual = getUserDataFromJSON json
expected = Just $ UD {
udAge = Nothing
, udNeedles = Nothing
, udTrunkLengthIncrementPerYear = Nothing
, udTrunkBranchLengthRatio = Nothing
, udTrunkBranchAngles = Nothing
, udTrunkGirth = Nothing
, udWhorlsPerYear = Nothing
, udWhorlSize = Nothing
, udBranchGirth = Nothing
, udBranchBranchLengthRatio = Nothing
, udBranchBranchLengthRatio2 = Nothing
-- , udBranchBranchAngle :: Angle Double
}
json = "\
\{}"
getUserDataFromJSON_Age = actual @?= expected
where actual = getUserDataFromJSON json
expected = Just $ UD {
udAge = Just 3
, udNeedles = Nothing
, udTrunkLengthIncrementPerYear = Nothing
, udTrunkBranchLengthRatio = Nothing
, udTrunkBranchAngles = Nothing
, udTrunkGirth = Nothing
, udWhorlsPerYear = Nothing
, udWhorlSize = Nothing
, udBranchGirth = Nothing
, udBranchBranchLengthRatio = Nothing
, udBranchBranchLengthRatio2 = Nothing
-- , udBranchBranchAngle :: Angle Double
}
json = "\
\{\"age\":3}"
getUserDataFromJSON_All = actual @?= expected
where actual = getUserDataFromJSON json
expected = Just $ UD {
udAge = Just 3
, udNeedles = Just False
, udTrunkLengthIncrementPerYear = Just 1.4
, udTrunkBranchLengthRatio = Just 0.6
, udTrunkBranchAngles = Just [0.698, 0.898, 1.31 , 0.967]
, udTrunkGirth = Just 5.0
, udWhorlsPerYear = Just 9
, udWhorlSize = Just 7
, udBranchGirth = Just 1.0
, udBranchBranchLengthRatio = Just 1.0
, udBranchBranchLengthRatio2 = Just 1.0
-- , udBranchBranchAngle :: Angle Double
}
json = "\
\{\"udTrunkGirth\":5,\"udWhorlsPerYear\":9,\"udTrunkBranchAngles\":[0.698,0.898,1.31,0.967],\"udTrunkBranchLengthRatio\":0.6,\"udBranchGirth\":1,\"udWhorlSize\":7,\"udBranchBranchLengthRatio\":1,\"udBranchBranchLengthRatio2\":1,\"age\":3,\"needles\":false,\"udTrunkLengthIncrementPerYear\":1.4}"
argsFromInput_NoChange = actual @?= expected
where actual = argsFromInput ud tp ap
expected = (tp, ap, False)
ud = UD {
udAge = Nothing
, udNeedles = Nothing
, udTrunkLengthIncrementPerYear = Nothing
, udTrunkBranchLengthRatio = Nothing
, udTrunkBranchAngles = Nothing
, udTrunkGirth = Nothing
, udWhorlsPerYear = Nothing
, udWhorlSize = Nothing
, udBranchGirth = Nothing
, udBranchBranchLengthRatio = Nothing
, udBranchBranchLengthRatio2 = Nothing
-- , udBranchBranchAngle :: Angle Double
}
tp = TreeParams {
tpTrunkLengthIncrementPerYear = 1.4
, tpTrunkBranchLengthRatio = 0.6
, tpTrunkBranchAngles = [0.698, 0.898, 1.31 , 0.967]
, tpTrunkGirth = 5.0
, tpWhorlsPerYear = 9
, tpWhorlSize = 7
, tpBranchGirth = 1.0
, tpBranchBranchLengthRatio = 1.0
, tpBranchBranchLengthRatio2 = 1.0
, tpBranchBranchAngle = 0.698
}
ap = AgeParams {
apAge = 3
, apTrunkBranchAngleIndex = 0
, apWhorlPhase = 0
}
argsFromInput_ChangeAll = actual @?= expected
where actual = argsFromInput ud tp ap
expected = (tp', ap', n)
ud = UD {
udAge = Just 5
, udNeedles = Just True
, udTrunkLengthIncrementPerYear = Just 1.5
, udTrunkBranchLengthRatio = Just 0.7
, udTrunkBranchAngles = Just [0.777]
, udTrunkGirth = Just 6.0
, udWhorlsPerYear = Just 10
, udWhorlSize = Just 5
, udBranchGirth = Just 1.2
, udBranchBranchLengthRatio = Just 1.3
, udBranchBranchLengthRatio2 = Just 1.4
-- , udBranchBranchAngle :: Angle Double
}
tp = TreeParams {
tpTrunkLengthIncrementPerYear = 1.4
, tpTrunkBranchLengthRatio = 0.6
, tpTrunkBranchAngles = [0.698, 0.898, 1.31 , 0.967]
, tpTrunkGirth = 5.0
, tpWhorlsPerYear = 9
, tpWhorlSize = 7
, tpBranchGirth = 1.0
, tpBranchBranchLengthRatio = 1.0
, tpBranchBranchLengthRatio2 = 1.0
, tpBranchBranchAngle = 0.698
}
ap = AgeParams {
apAge = 3
, apTrunkBranchAngleIndex = 0
, apWhorlPhase = 0
}
tp' = TreeParams {
tpTrunkLengthIncrementPerYear = 1.5
, tpTrunkBranchLengthRatio = 0.7
, tpTrunkBranchAngles = [0.777]
, tpTrunkGirth = 6.0
, tpWhorlsPerYear = 10
, tpWhorlSize = 5
, tpBranchGirth = 1.2
, tpBranchBranchLengthRatio = 1.3
, tpBranchBranchLengthRatio2 = 1.4
, tpBranchBranchAngle = 0.698
}
ap' = AgeParams {
apAge = 5
, apTrunkBranchAngleIndex = 0
, apWhorlPhase = 0
}
n = True
| bobgru/conifer | test/unit-tests/parserTests.hs | bsd-3-clause | 8,654 | 0 | 11 | 4,700 | 1,074 | 643 | 431 | 135 | 1 |
-- |
-- Module : Crypto.Internal.DeepSeq
-- License : BSD-style
-- Maintainer : Vincent Hanquez <[email protected]>
-- Stability : experimental
-- Portability : unknown
--
-- Simple abstraction module to allow compilation without deepseq
-- by defining our own NFData class if not compiling with deepseq
-- support.
--
{-# LANGUAGE CPP #-}
module Crypto.Internal.DeepSeq
( NFData(..)
) where
#ifdef WITH_DEEPSEQ_SUPPORT
import Control.DeepSeq
#else
import Data.Word
import Data.ByteArray
class NFData a where rnf :: a -> ()
instance NFData Word8 where rnf w = w `seq` ()
instance NFData Word16 where rnf w = w `seq` ()
instance NFData Word32 where rnf w = w `seq` ()
instance NFData Word64 where rnf w = w `seq` ()
instance NFData Bytes where rnf b = b `seq` ()
instance NFData ScrubbedBytes where rnf b = b `seq` ()
#endif
| nomeata/cryptonite | Crypto/Internal/DeepSeq.hs | bsd-3-clause | 849 | 0 | 5 | 163 | 36 | 29 | 7 | 12 | 0 |
{-# LANGUAGE EmptyDataDecls #-}
{-# LANGUAGE FlexibleContexts #-}
{-# LANGUAGE GADTs #-}
{-# LANGUAGE GeneralizedNewtypeDeriving #-}
{-# LANGUAGE MultiParamTypeClasses #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE QuasiQuotes #-}
{-# LANGUAGE TemplateHaskell #-}
{-# LANGUAGE TypeFamilies #-}
{-# LANGUAGE DeriveGeneric #-}
module Account.Types where
import Database.Persist
import Database.Persist.Sqlite
import Database.Persist.TH
import Data.Aeson
import GHC.Generics
{--
IGNORRE THE FOLLOWING TEMPLATE HASKELL BS
USING THIS IS BETTER THAN ACTUALLY WRITING DB STUFF
BUT IT'S REALLY UNIDIOMATIC
AND FRAGILE
--}
share [mkPersist sqlSettings, mkMigrate "migrateAll"] [persistLowerCase|
Account
name String
wins Int
losses Int
UniqueName name
deriving Show Generic
|]
instance ToJSON Account
instance FromJSON Account
| octopuscabbage/UltimateTicTacToeServer | src/Account/Types.hs | bsd-3-clause | 948 | 0 | 7 | 226 | 82 | 51 | 31 | 19 | 0 |
{-# LANGUAGE TemplateHaskell #-}
module AWS.CloudWatch.Types
where
import AWS.Lib.FromText
data Metric = Metric
{ metricDimensions :: [Dimension]
, metricName :: Text
, metricNameSpace :: Text
}
deriving (Show, Eq)
data Dimension = Dimension
{ dimensionName :: Text
, dimensionValue :: Text
}
deriving (Show, Eq)
type DimensionFilter = (Text, Text)
data Datapoint = Datapoint
{ datapointTimestamp :: UTCTime
, datapointSampleCount :: Maybe Double
, datapointUnit :: Text
, datapointMinimum :: Maybe Double
, datapointMaximum :: Maybe Double
, datapointSum :: Maybe Double
, datapointAverage :: Maybe Double
}
deriving (Show, Eq)
data Statistic
= StatisticAverage
| StatisticSum
| StatisticSampleCount
| StatisticMaximum
| StatisticMinimum
deriving (Show, Eq, Read)
deriveFromText "Statistic" ["Average", "Sum", "SampleCount", "Maximum", "Minimum"]
allStatistics :: [Statistic]
allStatistics =
[ StatisticAverage
, StatisticSum
, StatisticSampleCount
, StatisticMaximum
, StatisticMinimum
]
data AlarmNameSpec
= AlarmSpecNothing
| AlarmSpecNamePrefix Text
| AlarmSpecNames [Text]
deriving (Show, Eq)
data StateValue
= StateValueOk
| StateValueAlarm
| StateValueInsufficientData
deriving (Show, Eq, Read)
deriveFromText "StateValue" ["OK", "ALARM", "INSUFFICIENT_DATA"]
data ComparisonOperator
= GreaterThanOrEqualToThreshold
| GreaterThanThreshold
| LessThanThreshold
| LessThanOrEqualToThreshold
deriving (Show, Eq, Read)
deriveFromText "ComparisonOperator"
[ "GreaterThanOrEqualToThreshold"
, "GreaterThanThreshold"
, "LessThanThreshold"
, "LessThanOrEqualToThreshold"
]
data MetricAlarm = MetricAlarm
{ metricAlarmAlarmDescription :: Maybe Text
, metricAlarmStateUpdatedTimestamp :: UTCTime
, metricAlarmInsufficientDataActions :: [Text]
, metricAlarmStateReasonData :: Maybe Text
, metricAlarmAlarmArn :: Text
, metricAlarmConfigurationUpdatedTimestamp :: UTCTime
, metricAlarmAlarmName :: Text
, metricAlarmStateValue :: StateValue
, metricAlarmPeriod :: Int
, metricAlarmOKActions :: [Text]
, metricAlarmActionsEnabled :: Bool
, metricAlarmNamespace :: Text
, metricAlarmThreshold :: Double
, metricAlarmEvaluationPeriods :: Int
, metricAlarmStatistic :: Statistic
, metricAlarmAlarmActions :: [Text]
, metricAlarmUnit :: Maybe Text
, metricAlarmStateReason :: Maybe Text
, metricAlarmDimensions :: [Dimension]
, metricAlarmComparisonOperator :: ComparisonOperator
, metricAlarmMetricName :: Text
}
deriving (Show, Eq)
data PutMetricAlarmRequest = PutMetricAlarmRequest
{ putMetricAlarmActionsEnabled :: Maybe Bool
, putMetricAlarmAlarmActions :: [Text]
, putMetricAlarmAlarmDescription :: Maybe Text
, putMetricAlarmAlarmName :: Text
, putMetricAlarmComparisonOperator :: ComparisonOperator
, putMetricAlarmDimensions :: [Dimension]
, putMetricAlarmEvaluationPeriods :: Int
, putMetricAlarmInsufficientDataActions :: [Text]
, putMetricAlarmMetricName :: Text
, putMetricAlarmNamespace :: Text
, putMetricAlarmOKActions :: [Text]
, putMetricAlarmPeriod :: Int
, putMetricAlarmStatistic :: Statistic
, putMetricAlarmThreshold :: Double
, putMetricAlarmUnit :: Maybe Text
}
deriving (Show, Eq)
data AlarmHistory = AlarmHistory
{ alarmHistoryTimestamp :: UTCTime
, alarmHistoryHistoryItemType :: HistoryType
, alarmHistoryAlarmName :: Text
, alarmHistoryHistoryData :: Text
, alarmHistoryHistorySummary :: Text
}
deriving (Show, Eq)
data HistoryType
= HistoryTypeConfigurationUpdate
| HistoryTypeStateUpdate
| HistoryTypeAction
deriving (Show, Eq, Read)
deriveFromText "HistoryType" ["ConfigurationUpdate", "StateUpdate", "Action"]
data MetricDatum = MetricDatum
{ metricDatumDimensions :: [Dimension]
, metricDatumMetricName :: Text
, metricDatumTimestamp :: Maybe UTCTime
, metricDatumUnit :: Maybe Text
, metricDatumValue :: MetricDatumValue
}
deriving (Show, Eq)
data MetricDatumValue
= MetricDatumValue Double
| MetricDatumStatisticValues StatisticSet
deriving (Show, Eq)
data StatisticSet = StatisticSet
{ statisticSetMaximum :: Double
, statisticSetMinimum :: Double
, statisticSetSampleCount :: Double
, statisticSetSum :: Double
}
deriving (Show, Eq)
| IanConnolly/aws-sdk-fork | AWS/CloudWatch/Types.hs | bsd-3-clause | 4,515 | 0 | 9 | 915 | 912 | 550 | 362 | 129 | 1 |
#!/usr/bin/env runhaskell
import Distribution.Simple
main = defaultMain
| bcoppens/Lattices | Setup.hs | gpl-3.0 | 74 | 0 | 4 | 9 | 12 | 7 | 5 | 2 | 1 |
module Tests where
import qualified Text.Rudeify.Tests
main :: IO ()
main = Text.Rudeify.Tests.main
| AaronShiny/rudeify | tests/Tests.hs | gpl-3.0 | 102 | 0 | 6 | 15 | 30 | 19 | 11 | 4 | 1 |
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="ru-RU">
<title>Windows WebDrivers</title>
<maps>
<homeID>top</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>Contents</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>Index</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Search</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Favorites</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | secdec/zap-extensions | addOns/webdrivers/webdriverwindows/src/main/javahelp/org/zaproxy/zap/extension/webdriverwindows/resources/help_ru_RU/helpset_ru_RU.hs | apache-2.0 | 963 | 77 | 66 | 156 | 407 | 206 | 201 | -1 | -1 |
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="fil-PH">
<title>TLS Debug | ZAP Extension</title>
<maps>
<homeID>top</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>Mga Nilalaman</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>Indeks</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Paghahanap</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Mga paborito</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | veggiespam/zap-extensions | addOns/tlsdebug/src/main/javahelp/org/zaproxy/zap/extension/tlsdebug/resources/help_fil_PH/helpset_fil_PH.hs | apache-2.0 | 985 | 80 | 66 | 162 | 421 | 212 | 209 | -1 | -1 |
{-
Copyright 2010-2012 Cognimeta Inc.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in
compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is
distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express
or implied. See the License for the specific language governing permissions and limitations under the License.
-}
{-# LANGUAGE TypeFamilies #-}
module Database.Perdure.AllocCopy (
AllocCopy(..),
allocCopyBits,
module Database.Perdure.WriteBits
) where
import Prelude ()
import Cgm.Prelude
import Cgm.Data.Word
import Cgm.Data.Len
import Database.Perdure.WriteBits
import Cgm.System.Endian
class Endian w => AllocCopy w where
allocCopyBitsSkip :: (BitSrc d, SrcDestState d ~ RealWorld) => Len w Word -> d -> d -> ST RealWorld (STPrimArray RealWorld Pinned w)
allocCopyBits :: (BitSrc d, SrcDestState d ~ RealWorld, AllocCopy w) => d -> d -> IO (PrimArray Pinned w)
allocCopyBits start end = stToIO $ allocCopyBitsSkip 0 start end >>= unsafeFreezeSTPrimArray
instance AllocCopy Word32 where
allocCopyBitsSkip skip start end = onWordConv
(apply wordConv1 <$> allocCopyBitsSkip (retract wordLenB skip) start end)
(error "allocCopyBitsSkip for Word32 not implemented")
instance AllocCopy Word64 where
allocCopyBitsSkip skip start end = onWordConv
(error "allocCopyBitsSkip for Word64 not implemented")
(apply wordConv1 <$> allocCopyBitsSkip (retract wordLenB skip) start end)
instance AllocCopy Word where
-- Starts writing after the specified length 'skip', which can later be used to write a header.
allocCopyBitsSkip skip start end = do
wBuf <- mkArray $ coarsenLen (addedBits end start) + skip
_ <- copyBits end start (aligned $ CWordSeq wBuf skip) >>= padIncompleteWord
return wBuf
| bitemyapp/perdure | src/Database/Perdure/AllocCopy.hs | apache-2.0 | 2,173 | 0 | 13 | 518 | 406 | 208 | 198 | 28 | 1 |
{-# OPTIONS_GHC -fno-warn-orphans #-}
{-# LANGUAGE BangPatterns, MagicHash, CPP, TypeFamilies #-}
#if __GLASGOW_HASKELL__ >= 702
{-# LANGUAGE Trustworthy #-}
#endif
-- Using TemplateHaskell in text unconditionally is unacceptable, as
-- it's a GHC boot library. TemplateHaskellQuotes was added in 8.0, so
-- this would seem to be a problem. However, GHC's policy of only
-- needing to be able to compile itself from the last few releases
-- allows us to use full-fat TH on older versions, while using THQ for
-- GHC versions that may be used for bootstrapping.
#if __GLASGOW_HASKELL__ >= 800
{-# LANGUAGE TemplateHaskellQuotes #-}
#else
{-# LANGUAGE TemplateHaskell #-}
#endif
-- |
-- Module : Data.Text.Lazy
-- Copyright : (c) 2009, 2010, 2012 Bryan O'Sullivan
--
-- License : BSD-style
-- Maintainer : [email protected]
-- Portability : GHC
--
-- A time and space-efficient implementation of Unicode text using
-- lists of packed arrays.
--
-- /Note/: Read below the synopsis for important notes on the use of
-- this module.
--
-- The representation used by this module is suitable for high
-- performance use and for streaming large quantities of data. It
-- provides a means to manipulate a large body of text without
-- requiring that the entire content be resident in memory.
--
-- Some operations, such as 'concat', 'append', 'reverse' and 'cons',
-- have better time complexity than their "Data.Text" equivalents, due
-- to the underlying representation being a list of chunks. For other
-- operations, lazy 'Text's are usually within a few percent of strict
-- ones, but often with better heap usage if used in a streaming
-- fashion. For data larger than available memory, or if you have
-- tight memory constraints, this module will be the only option.
--
-- This module is intended to be imported @qualified@, to avoid name
-- clashes with "Prelude" functions. eg.
--
-- > import qualified Data.Text.Lazy as L
module Data.Text.Lazy
(
-- * Fusion
-- $fusion
-- * Acceptable data
-- $replacement
-- * Types
Text
-- * Creation and elimination
, pack
, unpack
, singleton
, empty
, fromChunks
, toChunks
, toStrict
, fromStrict
, foldrChunks
, foldlChunks
-- * Basic interface
, cons
, snoc
, append
, uncons
, unsnoc
, head
, last
, tail
, init
, null
, length
, compareLength
-- * Transformations
, map
, intercalate
, intersperse
, transpose
, reverse
, replace
-- ** Case conversion
-- $case
, toCaseFold
, toLower
, toUpper
, toTitle
-- ** Justification
, justifyLeft
, justifyRight
, center
-- * Folds
, foldl
, foldl'
, foldl1
, foldl1'
, foldr
, foldr1
-- ** Special folds
, concat
, concatMap
, any
, all
, maximum
, minimum
-- * Construction
-- ** Scans
, scanl
, scanl1
, scanr
, scanr1
-- ** Accumulating maps
, mapAccumL
, mapAccumR
-- ** Generation and unfolding
, repeat
, replicate
, cycle
, iterate
, unfoldr
, unfoldrN
-- * Substrings
-- ** Breaking strings
, take
, takeEnd
, drop
, dropEnd
, takeWhile
, takeWhileEnd
, dropWhile
, dropWhileEnd
, dropAround
, strip
, stripStart
, stripEnd
, splitAt
, span
, breakOn
, breakOnEnd
, break
, group
, groupBy
, inits
, tails
-- ** Breaking into many substrings
-- $split
, splitOn
, split
, chunksOf
-- , breakSubstring
-- ** Breaking into lines and words
, lines
, words
, unlines
, unwords
-- * Predicates
, isPrefixOf
, isSuffixOf
, isInfixOf
-- ** View patterns
, stripPrefix
, stripSuffix
, commonPrefixes
-- * Searching
, filter
, find
, elem
, breakOnAll
, partition
-- , findSubstring
-- * Indexing
, index
, count
-- * Zipping and unzipping
, zip
, zipWith
-- -* Ordered text
-- , sort
) where
import Prelude (Char, Bool(..), Maybe(..), String,
Eq(..), Ord(..), Ordering(..), Read(..), Show(..),
(&&), (||), (+), (-), (.), ($), (++),
error, flip, fmap, fromIntegral, not, otherwise, quot)
import qualified Prelude as P
import Control.DeepSeq (NFData(..))
import Data.Int (Int64)
import qualified Data.List as L
import Data.Char (isSpace)
import Data.Data (Data(gfoldl, toConstr, gunfold, dataTypeOf), constrIndex,
Constr, mkConstr, DataType, mkDataType, Fixity(Prefix))
import Data.Binary (Binary(get, put))
import Data.Monoid (Monoid(..))
#if MIN_VERSION_base(4,9,0)
import Data.Semigroup (Semigroup(..))
#endif
import Data.String (IsString(..))
import qualified Data.Text as T
import qualified Data.Text.Internal as T
import qualified Data.Text.Internal.Fusion.Common as S
import qualified Data.Text.Unsafe as T
import qualified Data.Text.Internal.Lazy.Fusion as S
import Data.Text.Internal.Fusion.Types (PairS(..))
import Data.Text.Internal.Lazy.Fusion (stream, unstream)
import Data.Text.Internal.Lazy (Text(..), chunk, empty, foldlChunks,
foldrChunks, smallChunkSize)
import Data.Text.Internal (firstf, safe, text)
import Data.Text.Lazy.Encoding (decodeUtf8', encodeUtf8)
import qualified Data.Text.Internal.Functions as F
import Data.Text.Internal.Lazy.Search (indices)
#if __GLASGOW_HASKELL__ >= 702
import qualified GHC.CString as GHC
#else
import qualified GHC.Base as GHC
#endif
#if MIN_VERSION_base(4,7,0)
import qualified GHC.Exts as Exts
#endif
import GHC.Prim (Addr#)
import qualified Language.Haskell.TH.Lib as TH
import qualified Language.Haskell.TH.Syntax as TH
#if MIN_VERSION_base(4,7,0)
import Text.Printf (PrintfArg, formatArg, formatString)
#endif
-- $fusion
--
-- Most of the functions in this module are subject to /fusion/,
-- meaning that a pipeline of such functions will usually allocate at
-- most one 'Text' value.
--
-- As an example, consider the following pipeline:
--
-- > import Data.Text.Lazy as T
-- > import Data.Text.Lazy.Encoding as E
-- > import Data.ByteString.Lazy (ByteString)
-- >
-- > countChars :: ByteString -> Int
-- > countChars = T.length . T.toUpper . E.decodeUtf8
--
-- From the type signatures involved, this looks like it should
-- allocate one 'ByteString' value, and two 'Text' values. However,
-- when a module is compiled with optimisation enabled under GHC, the
-- two intermediate 'Text' values will be optimised away, and the
-- function will be compiled down to a single loop over the source
-- 'ByteString'.
--
-- Functions that can be fused by the compiler are documented with the
-- phrase \"Subject to fusion\".
-- $replacement
--
-- A 'Text' value is a sequence of Unicode scalar values, as defined
-- in
-- <http://www.unicode.org/versions/Unicode5.2.0/ch03.pdf#page=35 §3.9, definition D76 of the Unicode 5.2 standard >.
-- As such, a 'Text' cannot contain values in the range U+D800 to
-- U+DFFF inclusive. Haskell implementations admit all Unicode code
-- points
-- (<http://www.unicode.org/versions/Unicode5.2.0/ch03.pdf#page=13 §3.4, definition D10 >)
-- as 'Char' values, including code points from this invalid range.
-- This means that there are some 'Char' values that are not valid
-- Unicode scalar values, and the functions in this module must handle
-- those cases.
--
-- Within this module, many functions construct a 'Text' from one or
-- more 'Char' values. Those functions will substitute 'Char' values
-- that are not valid Unicode scalar values with the replacement
-- character \"�\" (U+FFFD). Functions that perform this
-- inspection and replacement are documented with the phrase
-- \"Performs replacement on invalid scalar values\".
--
-- (One reason for this policy of replacement is that internally, a
-- 'Text' value is represented as packed UTF-16 data. Values in the
-- range U+D800 through U+DFFF are used by UTF-16 to denote surrogate
-- code points, and so cannot be represented. The functions replace
-- invalid scalar values, instead of dropping them, as a security
-- measure. For details, see
-- <http://unicode.org/reports/tr36/#Deletion_of_Noncharacters Unicode Technical Report 36, §3.5 >.)
-- $setup
-- >>> import Data.Text
-- >>> import qualified Data.Text as T
-- >>> :seti -XOverloadedStrings
equal :: Text -> Text -> Bool
equal Empty Empty = True
equal Empty _ = False
equal _ Empty = False
equal (Chunk a as) (Chunk b bs) =
case compare lenA lenB of
LT -> a == (T.takeWord16 lenA b) &&
as `equal` Chunk (T.dropWord16 lenA b) bs
EQ -> a == b && as `equal` bs
GT -> T.takeWord16 lenB a == b &&
Chunk (T.dropWord16 lenB a) as `equal` bs
where lenA = T.lengthWord16 a
lenB = T.lengthWord16 b
instance Eq Text where
(==) = equal
{-# INLINE (==) #-}
instance Ord Text where
compare = compareText
compareText :: Text -> Text -> Ordering
compareText Empty Empty = EQ
compareText Empty _ = LT
compareText _ Empty = GT
compareText (Chunk a0 as) (Chunk b0 bs) = outer a0 b0
where
outer ta@(T.Text arrA offA lenA) tb@(T.Text arrB offB lenB) = go 0 0
where
go !i !j
| i >= lenA = compareText as (chunk (T.Text arrB (offB+j) (lenB-j)) bs)
| j >= lenB = compareText (chunk (T.Text arrA (offA+i) (lenA-i)) as) bs
| a < b = LT
| a > b = GT
| otherwise = go (i+di) (j+dj)
where T.Iter a di = T.iter ta i
T.Iter b dj = T.iter tb j
instance Show Text where
showsPrec p ps r = showsPrec p (unpack ps) r
instance Read Text where
readsPrec p str = [(pack x,y) | (x,y) <- readsPrec p str]
#if MIN_VERSION_base(4,9,0)
-- | Non-orphan 'Semigroup' instance only defined for
-- @base-4.9.0.0@ and later; orphan instances for older GHCs are
-- provided by
-- the [semigroups](http://hackage.haskell.org/package/semigroups)
-- package
--
-- @since 1.2.2.0
instance Semigroup Text where
(<>) = append
#endif
instance Monoid Text where
mempty = empty
#if MIN_VERSION_base(4,9,0)
mappend = (<>) -- future-proof definition
#else
mappend = append
#endif
mconcat = concat
instance IsString Text where
fromString = pack
#if MIN_VERSION_base(4,7,0)
-- | @since 1.2.0.0
instance Exts.IsList Text where
type Item Text = Char
fromList = pack
toList = unpack
#endif
instance NFData Text where
rnf Empty = ()
rnf (Chunk _ ts) = rnf ts
-- | @since 1.2.1.0
instance Binary Text where
put t = put (encodeUtf8 t)
get = do
bs <- get
case decodeUtf8' bs of
P.Left exn -> P.fail (P.show exn)
P.Right a -> P.return a
-- | This instance preserves data abstraction at the cost of inefficiency.
-- We omit reflection services for the sake of data abstraction.
--
-- This instance was created by copying the updated behavior of
-- @"Data.Text".@'Data.Text.Text'
instance Data Text where
gfoldl f z txt = z pack `f` (unpack txt)
toConstr _ = packConstr
gunfold k z c = case constrIndex c of
1 -> k (z pack)
_ -> error "Data.Text.Lazy.Text.gunfold"
dataTypeOf _ = textDataType
-- | This instance has similar considerations to the 'Data' instance:
-- it preserves abstraction at the cost of inefficiency.
--
-- @since 1.2.4.0
instance TH.Lift Text where
lift = TH.appE (TH.varE 'pack) . TH.stringE . unpack
#if MIN_VERSION_template_haskell(2,17,0)
liftTyped = TH.unsafeCodeCoerce . TH.lift
#elif MIN_VERSION_template_haskell(2,16,0)
liftTyped = TH.unsafeTExpCoerce . TH.lift
#endif
#if MIN_VERSION_base(4,7,0)
-- | Only defined for @base-4.7.0.0@ and later
--
-- @since 1.2.2.0
instance PrintfArg Text where
formatArg txt = formatString $ unpack txt
#endif
packConstr :: Constr
packConstr = mkConstr textDataType "pack" [] Prefix
textDataType :: DataType
textDataType = mkDataType "Data.Text.Lazy.Text" [packConstr]
-- | /O(n)/ Convert a 'String' into a 'Text'.
--
-- Subject to fusion. Performs replacement on invalid scalar values.
pack :: String -> Text
pack = unstream . S.streamList . L.map safe
{-# INLINE [1] pack #-}
-- | /O(n)/ Convert a 'Text' into a 'String'.
-- Subject to fusion.
unpack :: Text -> String
unpack t = S.unstreamList (stream t)
{-# INLINE [1] unpack #-}
-- | /O(n)/ Convert a literal string into a Text.
unpackCString# :: Addr# -> Text
unpackCString# addr# = unstream (S.streamCString# addr#)
{-# NOINLINE unpackCString# #-}
{-# RULES "TEXT literal" forall a.
unstream (S.streamList (L.map safe (GHC.unpackCString# a)))
= unpackCString# a #-}
{-# RULES "TEXT literal UTF8" forall a.
unstream (S.streamList (L.map safe (GHC.unpackCStringUtf8# a)))
= unpackCString# a #-}
{-# RULES "LAZY TEXT empty literal"
unstream (S.streamList (L.map safe []))
= Empty #-}
{-# RULES "LAZY TEXT empty literal" forall a.
unstream (S.streamList (L.map safe [a]))
= Chunk (T.singleton a) Empty #-}
-- | /O(1)/ Convert a character into a Text. Subject to fusion.
-- Performs replacement on invalid scalar values.
singleton :: Char -> Text
singleton c = Chunk (T.singleton c) Empty
{-# INLINE [1] singleton #-}
{-# RULES
"LAZY TEXT singleton -> fused" [~1] forall c.
singleton c = unstream (S.singleton c)
"LAZY TEXT singleton -> unfused" [1] forall c.
unstream (S.singleton c) = singleton c
#-}
-- | /O(c)/ Convert a list of strict 'T.Text's into a lazy 'Text'.
fromChunks :: [T.Text] -> Text
fromChunks cs = L.foldr chunk Empty cs
-- | /O(n)/ Convert a lazy 'Text' into a list of strict 'T.Text's.
toChunks :: Text -> [T.Text]
toChunks cs = foldrChunks (:) [] cs
-- | /O(n)/ Convert a lazy 'Text' into a strict 'T.Text'.
toStrict :: Text -> T.Text
toStrict t = T.concat (toChunks t)
{-# INLINE [1] toStrict #-}
-- | /O(c)/ Convert a strict 'T.Text' into a lazy 'Text'.
fromStrict :: T.Text -> Text
fromStrict t = chunk t Empty
{-# INLINE [1] fromStrict #-}
-- -----------------------------------------------------------------------------
-- * Basic functions
-- | /O(1)/ Adds a character to the front of a 'Text'. Subject to fusion.
cons :: Char -> Text -> Text
cons c t = Chunk (T.singleton c) t
{-# INLINE [1] cons #-}
infixr 5 `cons`
{-# RULES
"LAZY TEXT cons -> fused" [~1] forall c t.
cons c t = unstream (S.cons c (stream t))
"LAZY TEXT cons -> unfused" [1] forall c t.
unstream (S.cons c (stream t)) = cons c t
#-}
-- | /O(n)/ Adds a character to the end of a 'Text'. This copies the
-- entire array in the process, unless fused. Subject to fusion.
snoc :: Text -> Char -> Text
snoc t c = foldrChunks Chunk (singleton c) t
{-# INLINE [1] snoc #-}
{-# RULES
"LAZY TEXT snoc -> fused" [~1] forall t c.
snoc t c = unstream (S.snoc (stream t) c)
"LAZY TEXT snoc -> unfused" [1] forall t c.
unstream (S.snoc (stream t) c) = snoc t c
#-}
-- | /O(n\/c)/ Appends one 'Text' to another. Subject to fusion.
append :: Text -> Text -> Text
append xs ys = foldrChunks Chunk ys xs
{-# INLINE [1] append #-}
{-# RULES
"LAZY TEXT append -> fused" [~1] forall t1 t2.
append t1 t2 = unstream (S.append (stream t1) (stream t2))
"LAZY TEXT append -> unfused" [1] forall t1 t2.
unstream (S.append (stream t1) (stream t2)) = append t1 t2
#-}
-- | /O(1)/ Returns the first character and rest of a 'Text', or
-- 'Nothing' if empty. Subject to fusion.
uncons :: Text -> Maybe (Char, Text)
uncons Empty = Nothing
uncons (Chunk t ts) = Just (T.unsafeHead t, ts')
where ts' | T.compareLength t 1 == EQ = ts
| otherwise = Chunk (T.unsafeTail t) ts
{-# INLINE uncons #-}
-- | /O(1)/ Returns the first character of a 'Text', which must be
-- non-empty. Subject to fusion.
head :: Text -> Char
head t = S.head (stream t)
{-# INLINE head #-}
-- | /O(1)/ Returns all characters after the head of a 'Text', which
-- must be non-empty. Subject to fusion.
tail :: Text -> Text
tail (Chunk t ts) = chunk (T.tail t) ts
tail Empty = emptyError "tail"
{-# INLINE [1] tail #-}
{-# RULES
"LAZY TEXT tail -> fused" [~1] forall t.
tail t = unstream (S.tail (stream t))
"LAZY TEXT tail -> unfused" [1] forall t.
unstream (S.tail (stream t)) = tail t
#-}
-- | /O(n\/c)/ Returns all but the last character of a 'Text', which must
-- be non-empty. Subject to fusion.
init :: Text -> Text
init (Chunk t0 ts0) = go t0 ts0
where go t (Chunk t' ts) = Chunk t (go t' ts)
go t Empty = chunk (T.init t) Empty
init Empty = emptyError "init"
{-# INLINE [1] init #-}
{-# RULES
"LAZY TEXT init -> fused" [~1] forall t.
init t = unstream (S.init (stream t))
"LAZY TEXT init -> unfused" [1] forall t.
unstream (S.init (stream t)) = init t
#-}
-- | /O(n\/c)/ Returns the 'init' and 'last' of a 'Text', or 'Nothing' if
-- empty.
--
-- * It is no faster than using 'init' and 'last'.
--
-- @since 1.2.3.0
unsnoc :: Text -> Maybe (Text, Char)
unsnoc Empty = Nothing
unsnoc ts@(Chunk _ _) = Just (init ts, last ts)
{-# INLINE unsnoc #-}
-- | /O(1)/ Tests whether a 'Text' is empty or not. Subject to
-- fusion.
null :: Text -> Bool
null Empty = True
null _ = False
{-# INLINE [1] null #-}
{-# RULES
"LAZY TEXT null -> fused" [~1] forall t.
null t = S.null (stream t)
"LAZY TEXT null -> unfused" [1] forall t.
S.null (stream t) = null t
#-}
-- | /O(1)/ Tests whether a 'Text' contains exactly one character.
-- Subject to fusion.
isSingleton :: Text -> Bool
isSingleton = S.isSingleton . stream
{-# INLINE isSingleton #-}
-- | /O(n\/c)/ Returns the last character of a 'Text', which must be
-- non-empty. Subject to fusion.
last :: Text -> Char
last Empty = emptyError "last"
last (Chunk t ts) = go t ts
where go _ (Chunk t' ts') = go t' ts'
go t' Empty = T.last t'
{-# INLINE [1] last #-}
{-# RULES
"LAZY TEXT last -> fused" [~1] forall t.
last t = S.last (stream t)
"LAZY TEXT last -> unfused" [1] forall t.
S.last (stream t) = last t
#-}
-- | /O(n)/ Returns the number of characters in a 'Text'.
-- Subject to fusion.
length :: Text -> Int64
length = foldlChunks go 0
where go l t = l + fromIntegral (T.length t)
{-# INLINE [1] length #-}
{-# RULES
"LAZY TEXT length -> fused" [~1] forall t.
length t = S.length (stream t)
"LAZY TEXT length -> unfused" [1] forall t.
S.length (stream t) = length t
#-}
-- | /O(n)/ Compare the count of characters in a 'Text' to a number.
-- Subject to fusion.
--
-- This function gives the same answer as comparing against the result
-- of 'length', but can short circuit if the count of characters is
-- greater than the number, and hence be more efficient.
compareLength :: Text -> Int64 -> Ordering
compareLength t n = S.compareLengthI (stream t) n
{-# INLINE [1] compareLength #-}
-- We don't apply those otherwise appealing length-to-compareLength
-- rewrite rules here, because they can change the strictness
-- properties of code.
-- | /O(n)/ 'map' @f@ @t@ is the 'Text' obtained by applying @f@ to
-- each element of @t@. Subject to fusion. Performs replacement on
-- invalid scalar values.
map :: (Char -> Char) -> Text -> Text
map f t = unstream (S.map (safe . f) (stream t))
{-# INLINE [1] map #-}
-- | /O(n)/ The 'intercalate' function takes a 'Text' and a list of
-- 'Text's and concatenates the list after interspersing the first
-- argument between each element of the list.
intercalate :: Text -> [Text] -> Text
intercalate t = concat . (F.intersperse t)
{-# INLINE intercalate #-}
-- | /O(n)/ The 'intersperse' function takes a character and places it
-- between the characters of a 'Text'. Subject to fusion. Performs
-- replacement on invalid scalar values.
intersperse :: Char -> Text -> Text
intersperse c t = unstream (S.intersperse (safe c) (stream t))
{-# INLINE intersperse #-}
-- | /O(n)/ Left-justify a string to the given length, using the
-- specified fill character on the right. Subject to fusion. Performs
-- replacement on invalid scalar values.
--
-- Examples:
--
-- > justifyLeft 7 'x' "foo" == "fooxxxx"
-- > justifyLeft 3 'x' "foobar" == "foobar"
justifyLeft :: Int64 -> Char -> Text -> Text
justifyLeft k c t
| len >= k = t
| otherwise = t `append` replicateChar (k-len) c
where len = length t
{-# INLINE [1] justifyLeft #-}
{-# RULES
"LAZY TEXT justifyLeft -> fused" [~1] forall k c t.
justifyLeft k c t = unstream (S.justifyLeftI k c (stream t))
"LAZY TEXT justifyLeft -> unfused" [1] forall k c t.
unstream (S.justifyLeftI k c (stream t)) = justifyLeft k c t
#-}
-- | /O(n)/ Right-justify a string to the given length, using the
-- specified fill character on the left. Performs replacement on
-- invalid scalar values.
--
-- Examples:
--
-- > justifyRight 7 'x' "bar" == "xxxxbar"
-- > justifyRight 3 'x' "foobar" == "foobar"
justifyRight :: Int64 -> Char -> Text -> Text
justifyRight k c t
| len >= k = t
| otherwise = replicateChar (k-len) c `append` t
where len = length t
{-# INLINE justifyRight #-}
-- | /O(n)/ Center a string to the given length, using the specified
-- fill character on either side. Performs replacement on invalid
-- scalar values.
--
-- Examples:
--
-- > center 8 'x' "HS" = "xxxHSxxx"
center :: Int64 -> Char -> Text -> Text
center k c t
| len >= k = t
| otherwise = replicateChar l c `append` t `append` replicateChar r c
where len = length t
d = k - len
r = d `quot` 2
l = d - r
{-# INLINE center #-}
-- | /O(n)/ The 'transpose' function transposes the rows and columns
-- of its 'Text' argument. Note that this function uses 'pack',
-- 'unpack', and the list version of transpose, and is thus not very
-- efficient.
transpose :: [Text] -> [Text]
transpose ts = L.map (\ss -> Chunk (T.pack ss) Empty)
(L.transpose (L.map unpack ts))
-- TODO: make this fast
-- | /O(n)/ 'reverse' @t@ returns the elements of @t@ in reverse order.
reverse :: Text -> Text
reverse = rev Empty
where rev a Empty = a
rev a (Chunk t ts) = rev (Chunk (T.reverse t) a) ts
-- | /O(m+n)/ Replace every non-overlapping occurrence of @needle@ in
-- @haystack@ with @replacement@.
--
-- This function behaves as though it was defined as follows:
--
-- @
-- replace needle replacement haystack =
-- 'intercalate' replacement ('splitOn' needle haystack)
-- @
--
-- As this suggests, each occurrence is replaced exactly once. So if
-- @needle@ occurs in @replacement@, that occurrence will /not/ itself
-- be replaced recursively:
--
-- > replace "oo" "foo" "oo" == "foo"
--
-- In cases where several instances of @needle@ overlap, only the
-- first one will be replaced:
--
-- > replace "ofo" "bar" "ofofo" == "barfo"
--
-- In (unlikely) bad cases, this function's time complexity degrades
-- towards /O(n*m)/.
replace :: Text
-- ^ @needle@ to search for. If this string is empty, an
-- error will occur.
-> Text
-- ^ @replacement@ to replace @needle@ with.
-> Text
-- ^ @haystack@ in which to search.
-> Text
replace s d = intercalate d . splitOn s
{-# INLINE replace #-}
-- ----------------------------------------------------------------------------
-- ** Case conversions (folds)
-- $case
--
-- With Unicode text, it is incorrect to use combinators like @map
-- toUpper@ to case convert each character of a string individually.
-- Instead, use the whole-string case conversion functions from this
-- module. For correctness in different writing systems, these
-- functions may map one input character to two or three output
-- characters.
-- | /O(n)/ Convert a string to folded case. Subject to fusion.
--
-- This function is mainly useful for performing caseless (or case
-- insensitive) string comparisons.
--
-- A string @x@ is a caseless match for a string @y@ if and only if:
--
-- @toCaseFold x == toCaseFold y@
--
-- The result string may be longer than the input string, and may
-- differ from applying 'toLower' to the input string. For instance,
-- the Armenian small ligature men now (U+FB13) is case folded to the
-- bigram men now (U+0574 U+0576), while the micro sign (U+00B5) is
-- case folded to the Greek small letter letter mu (U+03BC) instead of
-- itself.
toCaseFold :: Text -> Text
toCaseFold t = unstream (S.toCaseFold (stream t))
{-# INLINE toCaseFold #-}
-- | /O(n)/ Convert a string to lower case, using simple case
-- conversion. Subject to fusion.
--
-- The result string may be longer than the input string. For
-- instance, the Latin capital letter I with dot above (U+0130) maps
-- to the sequence Latin small letter i (U+0069) followed by combining
-- dot above (U+0307).
toLower :: Text -> Text
toLower t = unstream (S.toLower (stream t))
{-# INLINE toLower #-}
-- | /O(n)/ Convert a string to upper case, using simple case
-- conversion. Subject to fusion.
--
-- The result string may be longer than the input string. For
-- instance, the German eszett (U+00DF) maps to the two-letter
-- sequence SS.
toUpper :: Text -> Text
toUpper t = unstream (S.toUpper (stream t))
{-# INLINE toUpper #-}
-- | /O(n)/ Convert a string to title case, using simple case
-- conversion. Subject to fusion.
--
-- The first letter of the input is converted to title case, as is
-- every subsequent letter that immediately follows a non-letter.
-- Every letter that immediately follows another letter is converted
-- to lower case.
--
-- The result string may be longer than the input string. For example,
-- the Latin small ligature fl (U+FB02) is converted to the
-- sequence Latin capital letter F (U+0046) followed by Latin small
-- letter l (U+006C).
--
-- /Note/: this function does not take language or culture specific
-- rules into account. For instance, in English, different style
-- guides disagree on whether the book name \"The Hill of the Red
-- Fox\" is correctly title cased—but this function will
-- capitalize /every/ word.
--
-- @since 1.0.0.0
toTitle :: Text -> Text
toTitle t = unstream (S.toTitle (stream t))
{-# INLINE toTitle #-}
-- | /O(n)/ 'foldl', applied to a binary operator, a starting value
-- (typically the left-identity of the operator), and a 'Text',
-- reduces the 'Text' using the binary operator, from left to right.
-- Subject to fusion.
foldl :: (a -> Char -> a) -> a -> Text -> a
foldl f z t = S.foldl f z (stream t)
{-# INLINE foldl #-}
-- | /O(n)/ A strict version of 'foldl'.
-- Subject to fusion.
foldl' :: (a -> Char -> a) -> a -> Text -> a
foldl' f z t = S.foldl' f z (stream t)
{-# INLINE foldl' #-}
-- | /O(n)/ A variant of 'foldl' that has no starting value argument,
-- and thus must be applied to a non-empty 'Text'. Subject to fusion.
foldl1 :: (Char -> Char -> Char) -> Text -> Char
foldl1 f t = S.foldl1 f (stream t)
{-# INLINE foldl1 #-}
-- | /O(n)/ A strict version of 'foldl1'. Subject to fusion.
foldl1' :: (Char -> Char -> Char) -> Text -> Char
foldl1' f t = S.foldl1' f (stream t)
{-# INLINE foldl1' #-}
-- | /O(n)/ 'foldr', applied to a binary operator, a starting value
-- (typically the right-identity of the operator), and a 'Text',
-- reduces the 'Text' using the binary operator, from right to left.
-- Subject to fusion.
foldr :: (Char -> a -> a) -> a -> Text -> a
foldr f z t = S.foldr f z (stream t)
{-# INLINE foldr #-}
-- | /O(n)/ A variant of 'foldr' that has no starting value argument,
-- and thus must be applied to a non-empty 'Text'. Subject to
-- fusion.
foldr1 :: (Char -> Char -> Char) -> Text -> Char
foldr1 f t = S.foldr1 f (stream t)
{-# INLINE foldr1 #-}
-- | /O(n)/ Concatenate a list of 'Text's.
concat :: [Text] -> Text
concat = to
where
go Empty css = to css
go (Chunk c cs) css = Chunk c (go cs css)
to [] = Empty
to (cs:css) = go cs css
{-# INLINE concat #-}
-- | /O(n)/ Map a function over a 'Text' that results in a 'Text', and
-- concatenate the results.
concatMap :: (Char -> Text) -> Text -> Text
concatMap f = concat . foldr ((:) . f) []
{-# INLINE concatMap #-}
-- | /O(n)/ 'any' @p@ @t@ determines whether any character in the
-- 'Text' @t@ satisfies the predicate @p@. Subject to fusion.
any :: (Char -> Bool) -> Text -> Bool
any p t = S.any p (stream t)
{-# INLINE any #-}
-- | /O(n)/ 'all' @p@ @t@ determines whether all characters in the
-- 'Text' @t@ satisfy the predicate @p@. Subject to fusion.
all :: (Char -> Bool) -> Text -> Bool
all p t = S.all p (stream t)
{-# INLINE all #-}
-- | /O(n)/ 'maximum' returns the maximum value from a 'Text', which
-- must be non-empty. Subject to fusion.
maximum :: Text -> Char
maximum t = S.maximum (stream t)
{-# INLINE maximum #-}
-- | /O(n)/ 'minimum' returns the minimum value from a 'Text', which
-- must be non-empty. Subject to fusion.
minimum :: Text -> Char
minimum t = S.minimum (stream t)
{-# INLINE minimum #-}
-- | /O(n)/ 'scanl' is similar to 'foldl', but returns a list of
-- successive reduced values from the left. Subject to fusion.
-- Performs replacement on invalid scalar values.
--
-- > scanl f z [x1, x2, ...] == [z, z `f` x1, (z `f` x1) `f` x2, ...]
--
-- Note that
--
-- > last (scanl f z xs) == foldl f z xs.
scanl :: (Char -> Char -> Char) -> Char -> Text -> Text
scanl f z t = unstream (S.scanl g z (stream t))
where g a b = safe (f a b)
{-# INLINE scanl #-}
-- | /O(n)/ 'scanl1' is a variant of 'scanl' that has no starting
-- value argument. Performs replacement on invalid scalar values.
--
-- > scanl1 f [x1, x2, ...] == [x1, x1 `f` x2, ...]
scanl1 :: (Char -> Char -> Char) -> Text -> Text
scanl1 f t0 = case uncons t0 of
Nothing -> empty
Just (t,ts) -> scanl f t ts
{-# INLINE scanl1 #-}
-- | /O(n)/ 'scanr' is the right-to-left dual of 'scanl'. Performs
-- replacement on invalid scalar values.
--
-- > scanr f v == reverse . scanl (flip f) v . reverse
scanr :: (Char -> Char -> Char) -> Char -> Text -> Text
scanr f v = reverse . scanl g v . reverse
where g a b = safe (f b a)
-- | /O(n)/ 'scanr1' is a variant of 'scanr' that has no starting
-- value argument. Performs replacement on invalid scalar values.
scanr1 :: (Char -> Char -> Char) -> Text -> Text
scanr1 f t | null t = empty
| otherwise = scanr f (last t) (init t)
-- | /O(n)/ Like a combination of 'map' and 'foldl''. Applies a
-- function to each element of a 'Text', passing an accumulating
-- parameter from left to right, and returns a final 'Text'. Performs
-- replacement on invalid scalar values.
mapAccumL :: (a -> Char -> (a,Char)) -> a -> Text -> (a, Text)
mapAccumL f = go
where
go z (Chunk c cs) = (z'', Chunk c' cs')
where (z', c') = T.mapAccumL f z c
(z'', cs') = go z' cs
go z Empty = (z, Empty)
{-# INLINE mapAccumL #-}
-- | The 'mapAccumR' function behaves like a combination of 'map' and
-- a strict 'foldr'; it applies a function to each element of a
-- 'Text', passing an accumulating parameter from right to left, and
-- returning a final value of this accumulator together with the new
-- 'Text'. Performs replacement on invalid scalar values.
mapAccumR :: (a -> Char -> (a,Char)) -> a -> Text -> (a, Text)
mapAccumR f = go
where
go z (Chunk c cs) = (z'', Chunk c' cs')
where (z'', c') = T.mapAccumR f z' c
(z', cs') = go z cs
go z Empty = (z, Empty)
{-# INLINE mapAccumR #-}
-- | @'repeat' x@ is an infinite 'Text', with @x@ the value of every
-- element.
--
-- @since 1.2.0.5
repeat :: Char -> Text
repeat c = let t = Chunk (T.replicate smallChunkSize (T.singleton c)) t
in t
-- | /O(n*m)/ 'replicate' @n@ @t@ is a 'Text' consisting of the input
-- @t@ repeated @n@ times.
replicate :: Int64 -> Text -> Text
replicate n t
| null t || n <= 0 = empty
| isSingleton t = replicateChar n (head t)
| otherwise = concat (rep 0)
where rep !i | i >= n = []
| otherwise = t : rep (i+1)
{-# INLINE [1] replicate #-}
-- | 'cycle' ties a finite, non-empty 'Text' into a circular one, or
-- equivalently, the infinite repetition of the original 'Text'.
--
-- @since 1.2.0.5
cycle :: Text -> Text
cycle Empty = emptyError "cycle"
cycle t = let t' = foldrChunks Chunk t' t
in t'
-- | @'iterate' f x@ returns an infinite 'Text' of repeated applications
-- of @f@ to @x@:
--
-- > iterate f x == [x, f x, f (f x), ...]
--
-- @since 1.2.0.5
iterate :: (Char -> Char) -> Char -> Text
iterate f c = let t c' = Chunk (T.singleton c') (t (f c'))
in t c
-- | /O(n)/ 'replicateChar' @n@ @c@ is a 'Text' of length @n@ with @c@ the
-- value of every element. Subject to fusion.
replicateChar :: Int64 -> Char -> Text
replicateChar n c = unstream (S.replicateCharI n (safe c))
{-# INLINE replicateChar #-}
{-# RULES
"LAZY TEXT replicate/singleton -> replicateChar" [~1] forall n c.
replicate n (singleton c) = replicateChar n c
"LAZY TEXT replicate/unstream/singleton -> replicateChar" [~1] forall n c.
replicate n (unstream (S.singleton c)) = replicateChar n c
#-}
-- | /O(n)/, where @n@ is the length of the result. The 'unfoldr'
-- function is analogous to the List 'L.unfoldr'. 'unfoldr' builds a
-- 'Text' from a seed value. The function takes the element and
-- returns 'Nothing' if it is done producing the 'Text', otherwise
-- 'Just' @(a,b)@. In this case, @a@ is the next 'Char' in the
-- string, and @b@ is the seed value for further production.
-- Subject to fusion.
-- Performs replacement on invalid scalar values.
unfoldr :: (a -> Maybe (Char,a)) -> a -> Text
unfoldr f s = unstream (S.unfoldr (firstf safe . f) s)
{-# INLINE unfoldr #-}
-- | /O(n)/ Like 'unfoldr', 'unfoldrN' builds a 'Text' from a seed
-- value. However, the length of the result should be limited by the
-- first argument to 'unfoldrN'. This function is more efficient than
-- 'unfoldr' when the maximum length of the result is known and
-- correct, otherwise its performance is similar to 'unfoldr'.
-- Subject to fusion.
-- Performs replacement on invalid scalar values.
unfoldrN :: Int64 -> (a -> Maybe (Char,a)) -> a -> Text
unfoldrN n f s = unstream (S.unfoldrN n (firstf safe . f) s)
{-# INLINE unfoldrN #-}
-- | /O(n)/ 'take' @n@, applied to a 'Text', returns the prefix of the
-- 'Text' of length @n@, or the 'Text' itself if @n@ is greater than
-- the length of the Text. Subject to fusion.
take :: Int64 -> Text -> Text
take i _ | i <= 0 = Empty
take i t0 = take' i t0
where take' 0 _ = Empty
take' _ Empty = Empty
take' n (Chunk t ts)
| n < len = Chunk (T.take (fromIntegral n) t) Empty
| otherwise = Chunk t (take' (n - len) ts)
where len = fromIntegral (T.length t)
{-# INLINE [1] take #-}
{-# RULES
"LAZY TEXT take -> fused" [~1] forall n t.
take n t = unstream (S.take n (stream t))
"LAZY TEXT take -> unfused" [1] forall n t.
unstream (S.take n (stream t)) = take n t
#-}
-- | /O(n)/ 'takeEnd' @n@ @t@ returns the suffix remaining after
-- taking @n@ characters from the end of @t@.
--
-- Examples:
--
-- > takeEnd 3 "foobar" == "bar"
--
-- @since 1.1.1.0
takeEnd :: Int64 -> Text -> Text
takeEnd n t0
| n <= 0 = empty
| otherwise = takeChunk n empty . L.reverse . toChunks $ t0
where takeChunk _ acc [] = acc
takeChunk i acc (t:ts)
| i <= l = chunk (T.takeEnd (fromIntegral i) t) acc
| otherwise = takeChunk (i-l) (Chunk t acc) ts
where l = fromIntegral (T.length t)
-- | /O(n)/ 'drop' @n@, applied to a 'Text', returns the suffix of the
-- 'Text' after the first @n@ characters, or the empty 'Text' if @n@
-- is greater than the length of the 'Text'. Subject to fusion.
drop :: Int64 -> Text -> Text
drop i t0
| i <= 0 = t0
| otherwise = drop' i t0
where drop' 0 ts = ts
drop' _ Empty = Empty
drop' n (Chunk t ts)
| n < len = Chunk (T.drop (fromIntegral n) t) ts
| otherwise = drop' (n - len) ts
where len = fromIntegral (T.length t)
{-# INLINE [1] drop #-}
{-# RULES
"LAZY TEXT drop -> fused" [~1] forall n t.
drop n t = unstream (S.drop n (stream t))
"LAZY TEXT drop -> unfused" [1] forall n t.
unstream (S.drop n (stream t)) = drop n t
#-}
-- | /O(n)/ 'dropEnd' @n@ @t@ returns the prefix remaining after
-- dropping @n@ characters from the end of @t@.
--
-- Examples:
--
-- > dropEnd 3 "foobar" == "foo"
--
-- @since 1.1.1.0
dropEnd :: Int64 -> Text -> Text
dropEnd n t0
| n <= 0 = t0
| otherwise = dropChunk n . L.reverse . toChunks $ t0
where dropChunk _ [] = empty
dropChunk m (t:ts)
| m >= l = dropChunk (m-l) ts
| otherwise = fromChunks . L.reverse $
T.dropEnd (fromIntegral m) t : ts
where l = fromIntegral (T.length t)
-- | /O(n)/ 'dropWords' @n@ returns the suffix with @n@ 'Word16'
-- values dropped, or the empty 'Text' if @n@ is greater than the
-- number of 'Word16' values present.
dropWords :: Int64 -> Text -> Text
dropWords i t0
| i <= 0 = t0
| otherwise = drop' i t0
where drop' 0 ts = ts
drop' _ Empty = Empty
drop' n (Chunk (T.Text arr off len) ts)
| n < len' = chunk (text arr (off+n') (len-n')) ts
| otherwise = drop' (n - len') ts
where len' = fromIntegral len
n' = fromIntegral n
-- | /O(n)/ 'takeWhile', applied to a predicate @p@ and a 'Text',
-- returns the longest prefix (possibly empty) of elements that
-- satisfy @p@. Subject to fusion.
takeWhile :: (Char -> Bool) -> Text -> Text
takeWhile p t0 = takeWhile' t0
where takeWhile' Empty = Empty
takeWhile' (Chunk t ts) =
case T.findIndex (not . p) t of
Just n | n > 0 -> Chunk (T.take n t) Empty
| otherwise -> Empty
Nothing -> Chunk t (takeWhile' ts)
{-# INLINE [1] takeWhile #-}
{-# RULES
"LAZY TEXT takeWhile -> fused" [~1] forall p t.
takeWhile p t = unstream (S.takeWhile p (stream t))
"LAZY TEXT takeWhile -> unfused" [1] forall p t.
unstream (S.takeWhile p (stream t)) = takeWhile p t
#-}
-- | /O(n)/ 'takeWhileEnd', applied to a predicate @p@ and a 'Text',
-- returns the longest suffix (possibly empty) of elements that
-- satisfy @p@.
-- Examples:
--
-- > takeWhileEnd (=='o') "foo" == "oo"
--
-- @since 1.2.2.0
takeWhileEnd :: (Char -> Bool) -> Text -> Text
takeWhileEnd p = takeChunk empty . L.reverse . toChunks
where takeChunk acc [] = acc
takeChunk acc (t:ts)
| T.lengthWord16 t' < T.lengthWord16 t
= chunk t' acc
| otherwise = takeChunk (Chunk t' acc) ts
where t' = T.takeWhileEnd p t
{-# INLINE takeWhileEnd #-}
-- | /O(n)/ 'dropWhile' @p@ @t@ returns the suffix remaining after
-- 'takeWhile' @p@ @t@. Subject to fusion.
dropWhile :: (Char -> Bool) -> Text -> Text
dropWhile p t0 = dropWhile' t0
where dropWhile' Empty = Empty
dropWhile' (Chunk t ts) =
case T.findIndex (not . p) t of
Just n -> Chunk (T.drop n t) ts
Nothing -> dropWhile' ts
{-# INLINE [1] dropWhile #-}
{-# RULES
"LAZY TEXT dropWhile -> fused" [~1] forall p t.
dropWhile p t = unstream (S.dropWhile p (stream t))
"LAZY TEXT dropWhile -> unfused" [1] forall p t.
unstream (S.dropWhile p (stream t)) = dropWhile p t
#-}
-- | /O(n)/ 'dropWhileEnd' @p@ @t@ returns the prefix remaining after
-- dropping characters that satisfy the predicate @p@ from the end of
-- @t@.
--
-- Examples:
--
-- > dropWhileEnd (=='.') "foo..." == "foo"
dropWhileEnd :: (Char -> Bool) -> Text -> Text
dropWhileEnd p = go
where go Empty = Empty
go (Chunk t Empty) = if T.null t'
then Empty
else Chunk t' Empty
where t' = T.dropWhileEnd p t
go (Chunk t ts) = case go ts of
Empty -> go (Chunk t Empty)
ts' -> Chunk t ts'
{-# INLINE dropWhileEnd #-}
-- | /O(n)/ 'dropAround' @p@ @t@ returns the substring remaining after
-- dropping characters that satisfy the predicate @p@ from both the
-- beginning and end of @t@.
dropAround :: (Char -> Bool) -> Text -> Text
dropAround p = dropWhile p . dropWhileEnd p
{-# INLINE [1] dropAround #-}
-- | /O(n)/ Remove leading white space from a string. Equivalent to:
--
-- > dropWhile isSpace
stripStart :: Text -> Text
stripStart = dropWhile isSpace
{-# INLINE stripStart #-}
-- | /O(n)/ Remove trailing white space from a string. Equivalent to:
--
-- > dropWhileEnd isSpace
stripEnd :: Text -> Text
stripEnd = dropWhileEnd isSpace
{-# INLINE [1] stripEnd #-}
-- | /O(n)/ Remove leading and trailing white space from a string.
-- Equivalent to:
--
-- > dropAround isSpace
strip :: Text -> Text
strip = dropAround isSpace
{-# INLINE [1] strip #-}
-- | /O(n)/ 'splitAt' @n t@ returns a pair whose first element is a
-- prefix of @t@ of length @n@, and whose second is the remainder of
-- the string. It is equivalent to @('take' n t, 'drop' n t)@.
splitAt :: Int64 -> Text -> (Text, Text)
splitAt = loop
where loop _ Empty = (empty, empty)
loop n t | n <= 0 = (empty, t)
loop n (Chunk t ts)
| n < len = let (t',t'') = T.splitAt (fromIntegral n) t
in (Chunk t' Empty, Chunk t'' ts)
| otherwise = let (ts',ts'') = loop (n - len) ts
in (Chunk t ts', ts'')
where len = fromIntegral (T.length t)
-- | /O(n)/ 'splitAtWord' @n t@ returns a strict pair whose first
-- element is a prefix of @t@ whose chunks contain @n@ 'Word16'
-- values, and whose second is the remainder of the string.
splitAtWord :: Int64 -> Text -> PairS Text Text
splitAtWord _ Empty = empty :*: empty
splitAtWord x (Chunk c@(T.Text arr off len) cs)
| y >= len = let h :*: t = splitAtWord (x-fromIntegral len) cs
in Chunk c h :*: t
| otherwise = chunk (text arr off y) empty :*:
chunk (text arr (off+y) (len-y)) cs
where y = fromIntegral x
-- | /O(n+m)/ Find the first instance of @needle@ (which must be
-- non-'null') in @haystack@. The first element of the returned tuple
-- is the prefix of @haystack@ before @needle@ is matched. The second
-- is the remainder of @haystack@, starting with the match.
--
-- Examples:
--
-- > breakOn "::" "a::b::c" ==> ("a", "::b::c")
-- > breakOn "/" "foobar" ==> ("foobar", "")
--
-- Laws:
--
-- > append prefix match == haystack
-- > where (prefix, match) = breakOn needle haystack
--
-- If you need to break a string by a substring repeatedly (e.g. you
-- want to break on every instance of a substring), use 'breakOnAll'
-- instead, as it has lower startup overhead.
--
-- This function is strict in its first argument, and lazy in its
-- second.
--
-- In (unlikely) bad cases, this function's time complexity degrades
-- towards /O(n*m)/.
breakOn :: Text -> Text -> (Text, Text)
breakOn pat src
| null pat = emptyError "breakOn"
| otherwise = case indices pat src of
[] -> (src, empty)
(x:_) -> let h :*: t = splitAtWord x src
in (h, t)
-- | /O(n+m)/ Similar to 'breakOn', but searches from the end of the string.
--
-- The first element of the returned tuple is the prefix of @haystack@
-- up to and including the last match of @needle@. The second is the
-- remainder of @haystack@, following the match.
--
-- > breakOnEnd "::" "a::b::c" ==> ("a::b::", "c")
breakOnEnd :: Text -> Text -> (Text, Text)
breakOnEnd pat src = let (a,b) = breakOn (reverse pat) (reverse src)
in (reverse b, reverse a)
{-# INLINE breakOnEnd #-}
-- | /O(n+m)/ Find all non-overlapping instances of @needle@ in
-- @haystack@. Each element of the returned list consists of a pair:
--
-- * The entire string prior to the /k/th match (i.e. the prefix)
--
-- * The /k/th match, followed by the remainder of the string
--
-- Examples:
--
-- > breakOnAll "::" ""
-- > ==> []
-- > breakOnAll "/" "a/b/c/"
-- > ==> [("a", "/b/c/"), ("a/b", "/c/"), ("a/b/c", "/")]
--
-- This function is strict in its first argument, and lazy in its
-- second.
--
-- In (unlikely) bad cases, this function's time complexity degrades
-- towards /O(n*m)/.
--
-- The @needle@ parameter may not be empty.
breakOnAll :: Text -- ^ @needle@ to search for
-> Text -- ^ @haystack@ in which to search
-> [(Text, Text)]
breakOnAll pat src
| null pat = emptyError "breakOnAll"
| otherwise = go 0 empty src (indices pat src)
where
go !n p s (x:xs) = let h :*: t = splitAtWord (x-n) s
h' = append p h
in (h',t) : go x h' t xs
go _ _ _ _ = []
-- | /O(n)/ 'break' is like 'span', but the prefix returned is over
-- elements that fail the predicate @p@.
--
-- >>> T.break (=='c') "180cm"
-- ("180","cm")
break :: (Char -> Bool) -> Text -> (Text, Text)
break p t0 = break' t0
where break' Empty = (empty, empty)
break' c@(Chunk t ts) =
case T.findIndex p t of
Nothing -> let (ts', ts'') = break' ts
in (Chunk t ts', ts'')
Just n | n == 0 -> (Empty, c)
| otherwise -> let (a,b) = T.splitAt n t
in (Chunk a Empty, Chunk b ts)
-- | /O(n)/ 'span', applied to a predicate @p@ and text @t@, returns
-- a pair whose first element is the longest prefix (possibly empty)
-- of @t@ of elements that satisfy @p@, and whose second is the
-- remainder of the list.
--
-- >>> T.span (=='0') "000AB"
-- ("000","AB")
span :: (Char -> Bool) -> Text -> (Text, Text)
span p = break (not . p)
{-# INLINE span #-}
-- | The 'group' function takes a 'Text' and returns a list of 'Text's
-- such that the concatenation of the result is equal to the argument.
-- Moreover, each sublist in the result contains only equal elements.
-- For example,
--
-- > group "Mississippi" = ["M","i","ss","i","ss","i","pp","i"]
--
-- It is a special case of 'groupBy', which allows the programmer to
-- supply their own equality test.
group :: Text -> [Text]
group = groupBy (==)
{-# INLINE group #-}
-- | The 'groupBy' function is the non-overloaded version of 'group'.
groupBy :: (Char -> Char -> Bool) -> Text -> [Text]
groupBy _ Empty = []
groupBy eq (Chunk t ts) = cons x ys : groupBy eq zs
where (ys,zs) = span (eq x) xs
x = T.unsafeHead t
xs = chunk (T.unsafeTail t) ts
-- | /O(n)/ Return all initial segments of the given 'Text',
-- shortest first.
inits :: Text -> [Text]
inits = (Empty :) . inits'
where inits' Empty = []
inits' (Chunk t ts) = L.map (\t' -> Chunk t' Empty) (L.tail (T.inits t))
++ L.map (Chunk t) (inits' ts)
-- | /O(n)/ Return all final segments of the given 'Text', longest
-- first.
tails :: Text -> [Text]
tails Empty = Empty : []
tails ts@(Chunk t ts')
| T.length t == 1 = ts : tails ts'
| otherwise = ts : tails (Chunk (T.unsafeTail t) ts')
-- $split
--
-- Splitting functions in this library do not perform character-wise
-- copies to create substrings; they just construct new 'Text's that
-- are slices of the original.
-- | /O(m+n)/ Break a 'Text' into pieces separated by the first 'Text'
-- argument (which cannot be an empty string), consuming the
-- delimiter. An empty delimiter is invalid, and will cause an error
-- to be raised.
--
-- Examples:
--
-- > splitOn "\r\n" "a\r\nb\r\nd\r\ne" == ["a","b","d","e"]
-- > splitOn "aaa" "aaaXaaaXaaaXaaa" == ["","X","X","X",""]
-- > splitOn "x" "x" == ["",""]
--
-- and
--
-- > intercalate s . splitOn s == id
-- > splitOn (singleton c) == split (==c)
--
-- (Note: the string @s@ to split on above cannot be empty.)
--
-- This function is strict in its first argument, and lazy in its
-- second.
--
-- In (unlikely) bad cases, this function's time complexity degrades
-- towards /O(n*m)/.
splitOn :: Text
-- ^ String to split on. If this string is empty, an error
-- will occur.
-> Text
-- ^ Input text.
-> [Text]
splitOn pat src
| null pat = emptyError "splitOn"
| isSingleton pat = split (== head pat) src
| otherwise = go 0 (indices pat src) src
where
go _ [] cs = [cs]
go !i (x:xs) cs = let h :*: t = splitAtWord (x-i) cs
in h : go (x+l) xs (dropWords l t)
l = foldlChunks (\a (T.Text _ _ b) -> a + fromIntegral b) 0 pat
{-# INLINE [1] splitOn #-}
{-# RULES
"LAZY TEXT splitOn/singleton -> split/==" [~1] forall c t.
splitOn (singleton c) t = split (==c) t
#-}
-- | /O(n)/ Splits a 'Text' into components delimited by separators,
-- where the predicate returns True for a separator element. The
-- resulting components do not contain the separators. Two adjacent
-- separators result in an empty component in the output. eg.
--
-- > split (=='a') "aabbaca" == ["","","bb","c",""]
-- > split (=='a') [] == [""]
split :: (Char -> Bool) -> Text -> [Text]
split _ Empty = [Empty]
split p (Chunk t0 ts0) = comb [] (T.split p t0) ts0
where comb acc (s:[]) Empty = revChunks (s:acc) : []
comb acc (s:[]) (Chunk t ts) = comb (s:acc) (T.split p t) ts
comb acc (s:ss) ts = revChunks (s:acc) : comb [] ss ts
comb _ [] _ = impossibleError "split"
{-# INLINE split #-}
-- | /O(n)/ Splits a 'Text' into components of length @k@. The last
-- element may be shorter than the other chunks, depending on the
-- length of the input. Examples:
--
-- > chunksOf 3 "foobarbaz" == ["foo","bar","baz"]
-- > chunksOf 4 "haskell.org" == ["hask","ell.","org"]
chunksOf :: Int64 -> Text -> [Text]
chunksOf k = go
where
go t = case splitAt k t of
(a,b) | null a -> []
| otherwise -> a : go b
{-# INLINE chunksOf #-}
-- | /O(n)/ Breaks a 'Text' up into a list of 'Text's at
-- newline 'Char's. The resulting strings do not contain newlines.
lines :: Text -> [Text]
lines Empty = []
lines t = let (l,t') = break ((==) '\n') t
in l : if null t' then []
else lines (tail t')
-- | /O(n)/ Breaks a 'Text' up into a list of words, delimited by 'Char's
-- representing white space.
words :: Text -> [Text]
words = L.filter (not . null) . split isSpace
{-# INLINE words #-}
-- | /O(n)/ Joins lines, after appending a terminating newline to
-- each.
unlines :: [Text] -> Text
unlines = concat . L.map (`snoc` '\n')
{-# INLINE unlines #-}
-- | /O(n)/ Joins words using single space characters.
unwords :: [Text] -> Text
unwords = intercalate (singleton ' ')
{-# INLINE unwords #-}
-- | /O(n)/ The 'isPrefixOf' function takes two 'Text's and returns
-- 'True' iff the first is a prefix of the second. Subject to fusion.
isPrefixOf :: Text -> Text -> Bool
isPrefixOf Empty _ = True
isPrefixOf _ Empty = False
isPrefixOf (Chunk x xs) (Chunk y ys)
| lx == ly = x == y && isPrefixOf xs ys
| lx < ly = x == yh && isPrefixOf xs (Chunk yt ys)
| otherwise = xh == y && isPrefixOf (Chunk xt xs) ys
where (xh,xt) = T.splitAt ly x
(yh,yt) = T.splitAt lx y
lx = T.length x
ly = T.length y
{-# INLINE [1] isPrefixOf #-}
{-# RULES
"LAZY TEXT isPrefixOf -> fused" [~1] forall s t.
isPrefixOf s t = S.isPrefixOf (stream s) (stream t)
"LAZY TEXT isPrefixOf -> unfused" [1] forall s t.
S.isPrefixOf (stream s) (stream t) = isPrefixOf s t
#-}
-- | /O(n)/ The 'isSuffixOf' function takes two 'Text's and returns
-- 'True' iff the first is a suffix of the second.
isSuffixOf :: Text -> Text -> Bool
isSuffixOf x y = reverse x `isPrefixOf` reverse y
{-# INLINE isSuffixOf #-}
-- TODO: a better implementation
-- | /O(n+m)/ The 'isInfixOf' function takes two 'Text's and returns
-- 'True' iff the first is contained, wholly and intact, anywhere
-- within the second.
--
-- This function is strict in its first argument, and lazy in its
-- second.
--
-- In (unlikely) bad cases, this function's time complexity degrades
-- towards /O(n*m)/.
isInfixOf :: Text -> Text -> Bool
isInfixOf needle haystack
| null needle = True
| isSingleton needle = S.elem (head needle) . S.stream $ haystack
| otherwise = not . L.null . indices needle $ haystack
{-# INLINE [1] isInfixOf #-}
{-# RULES
"LAZY TEXT isInfixOf/singleton -> S.elem/S.stream" [~1] forall n h.
isInfixOf (singleton n) h = S.elem n (S.stream h)
#-}
-------------------------------------------------------------------------------
-- * View patterns
-- | /O(n)/ Return the suffix of the second string if its prefix
-- matches the entire first string.
--
-- Examples:
--
-- > stripPrefix "foo" "foobar" == Just "bar"
-- > stripPrefix "" "baz" == Just "baz"
-- > stripPrefix "foo" "quux" == Nothing
--
-- This is particularly useful with the @ViewPatterns@ extension to
-- GHC, as follows:
--
-- > {-# LANGUAGE ViewPatterns #-}
-- > import Data.Text.Lazy as T
-- >
-- > fnordLength :: Text -> Int
-- > fnordLength (stripPrefix "fnord" -> Just suf) = T.length suf
-- > fnordLength _ = -1
stripPrefix :: Text -> Text -> Maybe Text
stripPrefix p t
| null p = Just t
| otherwise = case commonPrefixes p t of
Just (_,c,r) | null c -> Just r
_ -> Nothing
-- | /O(n)/ Find the longest non-empty common prefix of two strings
-- and return it, along with the suffixes of each string at which they
-- no longer match.
--
-- If the strings do not have a common prefix or either one is empty,
-- this function returns 'Nothing'.
--
-- Examples:
--
-- > commonPrefixes "foobar" "fooquux" == Just ("foo","bar","quux")
-- > commonPrefixes "veeble" "fetzer" == Nothing
-- > commonPrefixes "" "baz" == Nothing
commonPrefixes :: Text -> Text -> Maybe (Text,Text,Text)
commonPrefixes Empty _ = Nothing
commonPrefixes _ Empty = Nothing
commonPrefixes a0 b0 = Just (go a0 b0 [])
where
go t0@(Chunk x xs) t1@(Chunk y ys) ps
= case T.commonPrefixes x y of
Just (p,a,b)
| T.null a -> go xs (chunk b ys) (p:ps)
| T.null b -> go (chunk a xs) ys (p:ps)
| otherwise -> (fromChunks (L.reverse (p:ps)),chunk a xs, chunk b ys)
Nothing -> (fromChunks (L.reverse ps),t0,t1)
go t0 t1 ps = (fromChunks (L.reverse ps),t0,t1)
-- | /O(n)/ Return the prefix of the second string if its suffix
-- matches the entire first string.
--
-- Examples:
--
-- > stripSuffix "bar" "foobar" == Just "foo"
-- > stripSuffix "" "baz" == Just "baz"
-- > stripSuffix "foo" "quux" == Nothing
--
-- This is particularly useful with the @ViewPatterns@ extension to
-- GHC, as follows:
--
-- > {-# LANGUAGE ViewPatterns #-}
-- > import Data.Text.Lazy as T
-- >
-- > quuxLength :: Text -> Int
-- > quuxLength (stripSuffix "quux" -> Just pre) = T.length pre
-- > quuxLength _ = -1
stripSuffix :: Text -> Text -> Maybe Text
stripSuffix p t = reverse `fmap` stripPrefix (reverse p) (reverse t)
-- | /O(n)/ 'filter', applied to a predicate and a 'Text',
-- returns a 'Text' containing those characters that satisfy the
-- predicate.
filter :: (Char -> Bool) -> Text -> Text
filter p t = unstream (S.filter p (stream t))
{-# INLINE filter #-}
-- | /O(n)/ The 'find' function takes a predicate and a 'Text', and
-- returns the first element in matching the predicate, or 'Nothing'
-- if there is no such element. Subject to fusion.
find :: (Char -> Bool) -> Text -> Maybe Char
find p t = S.findBy p (stream t)
{-# INLINE find #-}
-- | /O(n)/ The 'elem' function takes a character and a 'Text', and
-- returns 'True' if the element is found in the given 'Text', or
-- 'False' otherwise.
elem :: Char -> Text -> Bool
elem c t = S.any (== c) (stream t)
{-# INLINE elem #-}
-- | /O(n)/ The 'partition' function takes a predicate and a 'Text',
-- and returns the pair of 'Text's with elements which do and do not
-- satisfy the predicate, respectively; i.e.
--
-- > partition p t == (filter p t, filter (not . p) t)
partition :: (Char -> Bool) -> Text -> (Text, Text)
partition p t = (filter p t, filter (not . p) t)
{-# INLINE partition #-}
-- | /O(n)/ 'Text' index (subscript) operator, starting from 0.
-- Subject to fusion.
index :: Text -> Int64 -> Char
index t n = S.index (stream t) n
{-# INLINE index #-}
-- | /O(n+m)/ The 'count' function returns the number of times the
-- query string appears in the given 'Text'. An empty query string is
-- invalid, and will cause an error to be raised.
--
-- In (unlikely) bad cases, this function's time complexity degrades
-- towards /O(n*m)/.
count :: Text -> Text -> Int64
count pat src
| null pat = emptyError "count"
| otherwise = go 0 (indices pat src)
where go !n [] = n
go !n (_:xs) = go (n+1) xs
{-# INLINE [1] count #-}
{-# RULES
"LAZY TEXT count/singleton -> countChar" [~1] forall c t.
count (singleton c) t = countChar c t
#-}
-- | /O(n)/ The 'countChar' function returns the number of times the
-- query element appears in the given 'Text'. Subject to fusion.
countChar :: Char -> Text -> Int64
countChar c t = S.countChar c (stream t)
-- | /O(n)/ 'zip' takes two 'Text's and returns a list of
-- corresponding pairs of bytes. If one input 'Text' is short,
-- excess elements of the longer 'Text' are discarded. This is
-- equivalent to a pair of 'unpack' operations.
zip :: Text -> Text -> [(Char,Char)]
zip a b = S.unstreamList $ S.zipWith (,) (stream a) (stream b)
{-# INLINE [0] zip #-}
-- | /O(n)/ 'zipWith' generalises 'zip' by zipping with the function
-- given as the first argument, instead of a tupling function.
-- Performs replacement on invalid scalar values.
zipWith :: (Char -> Char -> Char) -> Text -> Text -> Text
zipWith f t1 t2 = unstream (S.zipWith g (stream t1) (stream t2))
where g a b = safe (f a b)
{-# INLINE [0] zipWith #-}
revChunks :: [T.Text] -> Text
revChunks = L.foldl' (flip chunk) Empty
emptyError :: String -> a
emptyError fun = P.error ("Data.Text.Lazy." ++ fun ++ ": empty input")
impossibleError :: String -> a
impossibleError fun = P.error ("Data.Text.Lazy." ++ fun ++ ": impossible case")
| bgamari/text | src/Data/Text/Lazy.hs | bsd-2-clause | 58,598 | 0 | 16 | 14,073 | 10,830 | 5,993 | 4,837 | 820 | 5 |
{-# LANGUAGE TypeSynonymInstances #-}
module Foundation.Math.Trigonometry
( Trigonometry(..)
) where
import Basement.Compat.Base
import qualified Prelude
-- | Method to support basic trigonometric functions
class Trigonometry a where
-- | the famous pi value
pi :: a
-- | sine
sin :: a -> a
-- | cosine
cos :: a -> a
-- | tan
tan :: a -> a
-- | sine-1
asin :: a -> a
-- | cosine-1
acos :: a -> a
-- | tangent-1
atan :: a -> a
-- | hyperbolic sine
sinh :: a -> a
-- | hyperbolic cosine
cosh :: a -> a
-- | hyperbolic tangent
tanh :: a -> a
-- | hyperbolic sine-1
asinh :: a -> a
-- | hyperbolic cosine-1
acosh :: a -> a
-- | hyperbolic tangent-1
atanh :: a -> a
instance Trigonometry Float where
pi = Prelude.pi
sin = Prelude.sin
cos = Prelude.cos
tan = Prelude.tan
asin = Prelude.asin
acos = Prelude.acos
atan = Prelude.atan
sinh = Prelude.sinh
cosh = Prelude.cosh
tanh = Prelude.tanh
asinh = Prelude.asinh
acosh = Prelude.acosh
atanh = Prelude.atanh
instance Trigonometry Double where
pi = Prelude.pi
sin = Prelude.sin
cos = Prelude.cos
tan = Prelude.tan
asin = Prelude.asin
acos = Prelude.acos
atan = Prelude.atan
sinh = Prelude.sinh
cosh = Prelude.cosh
tanh = Prelude.tanh
asinh = Prelude.asinh
acosh = Prelude.acosh
atanh = Prelude.atanh
| vincenthz/hs-foundation | foundation/Foundation/Math/Trigonometry.hs | bsd-3-clause | 1,487 | 0 | 7 | 469 | 361 | 212 | 149 | 47 | 0 |
{-# LANGUAGE PatternGuards, CPP, ForeignFunctionInterface #-}
-----------------------------------------------------------------------------
--
-- (c) The University of Glasgow 2004-2009.
--
-- Package management tool
--
-----------------------------------------------------------------------------
module Main (main) where
import Distribution.InstalledPackageInfo.Binary()
import qualified Distribution.Simple.PackageIndex as PackageIndex
import Distribution.ModuleName hiding (main)
import Distribution.InstalledPackageInfo
import Distribution.Compat.ReadP
import Distribution.ParseUtils
import Distribution.Package hiding (depends)
import Distribution.Text
import Distribution.Version
import System.FilePath as FilePath
import qualified System.FilePath.Posix as FilePath.Posix
import System.Process
import System.Directory ( getAppUserDataDirectory, createDirectoryIfMissing,
getModificationTime )
import Text.Printf
import Prelude
import System.Console.GetOpt
import qualified Control.Exception as Exception
import Data.Maybe
import Data.Char ( isSpace, toLower )
import Data.Ord (comparing)
import Control.Applicative (Applicative(..))
import Control.Monad
import System.Directory ( doesDirectoryExist, getDirectoryContents,
doesFileExist, renameFile, removeFile,
getCurrentDirectory )
import System.Exit ( exitWith, ExitCode(..) )
import System.Environment ( getArgs, getProgName, getEnv )
import System.IO
import System.IO.Error
import Data.List
import Control.Concurrent
import qualified Data.ByteString.Lazy as B
import qualified Data.Binary as Bin
import qualified Data.Binary.Get as Bin
-- Haste-specific
import Haste.Environment
import Haste.Version
import System.Info (os)
import qualified Control.Shell as Sh
#if defined(mingw32_HOST_OS)
-- mingw32 needs these for getExecDir
import Foreign
import Foreign.C
#endif
#ifdef mingw32_HOST_OS
import GHC.ConsoleHandler
#else
import System.Posix hiding (fdToHandle)
#endif
#if defined(GLOB)
import qualified System.Info(os)
#endif
#if !defined(mingw32_HOST_OS) && !defined(BOOTSTRAPPING)
import System.Console.Terminfo as Terminfo
#endif
#ifdef mingw32_HOST_OS
# if defined(i386_HOST_ARCH)
# define WINDOWS_CCONV stdcall
# elif defined(x86_64_HOST_ARCH)
# define WINDOWS_CCONV ccall
# else
# error Unknown mingw32 arch
# endif
#endif
-- -----------------------------------------------------------------------------
-- Entry point
main :: IO ()
main = do
args <- getArgs
case args of
["relocate", pkg] -> do
Sh.shell (relocate packages pkg) >> exitWith ExitSuccess
_ ->
return ()
case getOpt Permute (flags ++ deprecFlags) args of
(cli,_,[]) | FlagHelp `elem` cli -> do
prog <- getProgramName
bye (usageInfo (usageHeader prog) flags)
(cli,_,[]) | FlagVersion `elem` cli ->
bye ourCopyright
(cli,nonopts,[]) ->
case getVerbosity Normal cli of
Right v -> runit v cli nonopts
Left err -> die err
(_,_,errors) -> do
prog <- getProgramName
die (concat errors ++ shortUsage prog)
where
packages = ["--global-package-db=" ++ pkgSysDir,
"--package-db=" ++ pkgSysDir,
"--package-db=" ++ pkgUserDir]
-- -----------------------------------------------------------------------------
-- Command-line syntax
data Flag
= FlagUser
| FlagGlobal
| FlagHelp
| FlagVersion
| FlagConfig FilePath
| FlagGlobalConfig FilePath
| FlagForce
| FlagForceFiles
| FlagAutoGHCiLibs
| FlagExpandEnvVars
| FlagExpandPkgroot
| FlagNoExpandPkgroot
| FlagSimpleOutput
| FlagNamesOnly
| FlagIgnoreCase
| FlagNoUserDb
| FlagVerbosity (Maybe String)
deriving Eq
flags :: [OptDescr Flag]
flags = [
Option [] ["user"] (NoArg FlagUser)
"use the current user's package database",
Option [] ["global"] (NoArg FlagGlobal)
"use the global package database",
Option ['f'] ["package-db"] (ReqArg FlagConfig "FILE/DIR")
"use the specified package database",
Option [] ["package-conf"] (ReqArg FlagConfig "FILE/DIR")
"use the specified package database (DEPRECATED)",
Option [] ["global-package-db"] (ReqArg FlagGlobalConfig "DIR")
"location of the global package database",
Option [] ["no-user-package-db"] (NoArg FlagNoUserDb)
"never read the user package database",
Option [] ["no-user-package-conf"] (NoArg FlagNoUserDb)
"never read the user package database (DEPRECATED)",
Option [] ["force"] (NoArg FlagForce)
"ignore missing dependencies, directories, and libraries",
Option [] ["force-files"] (NoArg FlagForceFiles)
"ignore missing directories and libraries only",
Option ['g'] ["auto-ghci-libs"] (NoArg FlagAutoGHCiLibs)
"automatically build libs for GHCi (with register)",
Option [] ["expand-env-vars"] (NoArg FlagExpandEnvVars)
"expand environment variables (${name}-style) in input package descriptions",
Option [] ["expand-pkgroot"] (NoArg FlagExpandPkgroot)
"expand ${pkgroot}-relative paths to absolute in output package descriptions",
Option [] ["no-expand-pkgroot"] (NoArg FlagNoExpandPkgroot)
"preserve ${pkgroot}-relative paths in output package descriptions",
Option ['?'] ["help"] (NoArg FlagHelp)
"display this help and exit",
Option ['V'] ["version"] (NoArg FlagVersion)
"output version information and exit",
Option [] ["simple-output"] (NoArg FlagSimpleOutput)
"print output in easy-to-parse format for some commands",
Option [] ["names-only"] (NoArg FlagNamesOnly)
"only print package names, not versions; can only be used with list --simple-output",
Option [] ["ignore-case"] (NoArg FlagIgnoreCase)
"ignore case for substring matching",
Option ['v'] ["verbose"] (OptArg FlagVerbosity "Verbosity")
"verbosity level (0-2, default 1)"
]
data Verbosity = Silent | Normal | Verbose
deriving (Show, Eq, Ord)
getVerbosity :: Verbosity -> [Flag] -> Either String Verbosity
getVerbosity v [] = Right v
getVerbosity _ (FlagVerbosity Nothing : fs) = getVerbosity Verbose fs
getVerbosity _ (FlagVerbosity (Just "0") : fs) = getVerbosity Silent fs
getVerbosity _ (FlagVerbosity (Just "1") : fs) = getVerbosity Normal fs
getVerbosity _ (FlagVerbosity (Just "2") : fs) = getVerbosity Verbose fs
getVerbosity _ (FlagVerbosity v : _) = Left ("Bad verbosity: " ++ show v)
getVerbosity v (_ : fs) = getVerbosity v fs
deprecFlags :: [OptDescr Flag]
deprecFlags = [
-- put deprecated flags here
]
ourCopyright :: String
ourCopyright = "Haste package manager version " ++ showVersion ghcVersion ++ "\n"
shortUsage :: String -> String
shortUsage prog = "For usage information see '" ++ prog ++ " --help'."
usageHeader :: String -> String
usageHeader prog = substProg prog $
"Usage:\n" ++
" $p init {path}\n" ++
" Create and initialise a package database at the location {path}.\n" ++
" Packages can be registered in the new database using the register\n" ++
" command with --package-db={path}. To use the new database with GHC,\n" ++
" use GHC's -package-db flag.\n" ++
"\n" ++
" $p register {filename | -}\n" ++
" Register the package using the specified installed package\n" ++
" description. The syntax for the latter is given in the $p\n" ++
" documentation. The input file should be encoded in UTF-8.\n" ++
"\n" ++
" $p update {filename | -}\n" ++
" Register the package, overwriting any other package with the\n" ++
" same name. The input file should be encoded in UTF-8.\n" ++
"\n" ++
" $p unregister {pkg-id}\n" ++
" Unregister the specified package.\n" ++
"\n" ++
" $p expose {pkg-id}\n" ++
" Expose the specified package.\n" ++
"\n" ++
" $p hide {pkg-id}\n" ++
" Hide the specified package.\n" ++
"\n" ++
" $p trust {pkg-id}\n" ++
" Trust the specified package.\n" ++
"\n" ++
" $p distrust {pkg-id}\n" ++
" Distrust the specified package.\n" ++
"\n" ++
" $p list [pkg]\n" ++
" List registered packages in the global database, and also the\n" ++
" user database if --user is given. If a package name is given\n" ++
" all the registered versions will be listed in ascending order.\n" ++
" Accepts the --simple-output flag.\n" ++
"\n" ++
" $p dot\n" ++
" Generate a graph of the package dependencies in a form suitable\n" ++
" for input for the graphviz tools. For example, to generate a PDF" ++
" of the dependency graph: ghc-pkg dot | tred | dot -Tpdf >pkgs.pdf" ++
"\n" ++
" $p find-module {module}\n" ++
" List registered packages exposing module {module} in the global\n" ++
" database, and also the user database if --user is given.\n" ++
" All the registered versions will be listed in ascending order.\n" ++
" Accepts the --simple-output flag.\n" ++
"\n" ++
" $p latest {pkg-id}\n" ++
" Prints the highest registered version of a package.\n" ++
"\n" ++
" $p check\n" ++
" Check the consistency of package dependencies and list broken packages.\n" ++
" Accepts the --simple-output flag.\n" ++
"\n" ++
" $p describe {pkg}\n" ++
" Give the registered description for the specified package. The\n" ++
" description is returned in precisely the syntax required by $p\n" ++
" register.\n" ++
"\n" ++
" $p field {pkg} {field}\n" ++
" Extract the specified field of the package description for the\n" ++
" specified package. Accepts comma-separated multiple fields.\n" ++
"\n" ++
" $p dump\n" ++
" Dump the registered description for every package. This is like\n" ++
" \"ghc-pkg describe '*'\", except that it is intended to be used\n" ++
" by tools that parse the results, rather than humans. The output is\n" ++
" always encoded in UTF-8, regardless of the current locale.\n" ++
"\n" ++
" $p recache\n" ++
" Regenerate the package database cache. This command should only be\n" ++
" necessary if you added a package to the database by dropping a file\n" ++
" into the database directory manually. By default, the global DB\n" ++
" is recached; to recache a different DB use --user or --package-db\n" ++
" as appropriate.\n" ++
"\n" ++
" Substring matching is supported for {module} in find-module and\n" ++
" for {pkg} in list, describe, and field, where a '*' indicates\n" ++
" open substring ends (prefix*, *suffix, *infix*).\n" ++
"\n" ++
" When asked to modify a database (register, unregister, update,\n"++
" hide, expose, and also check), ghc-pkg modifies the global database by\n"++
" default. Specifying --user causes it to act on the user database,\n"++
" or --package-db can be used to act on another database\n"++
" entirely. When multiple of these options are given, the rightmost\n"++
" one is used as the database to act upon.\n"++
"\n"++
" Commands that query the package database (list, tree, latest, describe,\n"++
" field) operate on the list of databases specified by the flags\n"++
" --user, --global, and --package-db. If none of these flags are\n"++
" given, the default is --global --user.\n"++
"\n" ++
" The following optional flags are also accepted:\n"
substProg :: String -> String -> String
substProg _ [] = []
substProg prog ('$':'p':xs) = prog ++ substProg prog xs
substProg prog (c:xs) = c : substProg prog xs
-- -----------------------------------------------------------------------------
-- Do the business
data Force = NoForce | ForceFiles | ForceAll | CannotForce
deriving (Eq,Ord)
data PackageArg = Id PackageIdentifier | Substring String (String->Bool)
runit :: Verbosity -> [Flag] -> [String] -> IO ()
runit verbosity cli nonopts = do
installSignalHandlers -- catch ^C and clean up
prog <- getProgramName
let
force
| FlagForce `elem` cli = ForceAll
| FlagForceFiles `elem` cli = ForceFiles
| otherwise = NoForce
auto_ghci_libs = FlagAutoGHCiLibs `elem` cli
expand_env_vars= FlagExpandEnvVars `elem` cli
mexpand_pkgroot= foldl' accumExpandPkgroot Nothing cli
where accumExpandPkgroot _ FlagExpandPkgroot = Just True
accumExpandPkgroot _ FlagNoExpandPkgroot = Just False
accumExpandPkgroot x _ = x
splitFields fields = unfoldr splitComma (',':fields)
where splitComma "" = Nothing
splitComma fs = Just $ break (==',') (tail fs)
substringCheck :: String -> Maybe (String -> Bool)
substringCheck "" = Nothing
substringCheck "*" = Just (const True)
substringCheck [_] = Nothing
substringCheck (h:t) =
case (h, init t, last t) of
('*',s,'*') -> Just (isInfixOf (f s) . f)
('*',_, _ ) -> Just (isSuffixOf (f t) . f)
( _ ,s,'*') -> Just (isPrefixOf (f (h:s)) . f)
_ -> Nothing
where f | FlagIgnoreCase `elem` cli = map toLower
| otherwise = id
#if defined(GLOB)
glob x | System.Info.os=="mingw32" = do
-- glob echoes its argument, after win32 filename globbing
(_,o,_,_) <- runInteractiveCommand ("glob "++x)
txt <- hGetContents o
return (read txt)
glob x | otherwise = return [x]
#endif
--
-- first, parse the command
case nonopts of
#if defined(GLOB)
-- dummy command to demonstrate usage and permit testing
-- without messing things up; use glob to selectively enable
-- windows filename globbing for file parameters
-- register, update, FlagGlobalConfig, FlagConfig; others?
["glob", filename] -> do
print filename
glob filename >>= print
#endif
["init", filename] ->
initPackageDB filename verbosity cli
["register", filename] ->
registerPackage filename verbosity cli
auto_ghci_libs expand_env_vars False force
["update", filename] ->
registerPackage filename verbosity cli
auto_ghci_libs expand_env_vars True force
["unregister", pkgid_str] -> do
pkgid <- readGlobPkgId pkgid_str
unregisterPackage pkgid verbosity cli force
["expose", pkgid_str] -> do
pkgid <- readGlobPkgId pkgid_str
exposePackage pkgid verbosity cli force
["hide", pkgid_str] -> do
pkgid <- readGlobPkgId pkgid_str
hidePackage pkgid verbosity cli force
["trust", pkgid_str] -> do
pkgid <- readGlobPkgId pkgid_str
trustPackage pkgid verbosity cli force
["distrust", pkgid_str] -> do
pkgid <- readGlobPkgId pkgid_str
distrustPackage pkgid verbosity cli force
["list"] -> do
listPackages verbosity cli Nothing Nothing
["list", pkgid_str] ->
case substringCheck pkgid_str of
Nothing -> do pkgid <- readGlobPkgId pkgid_str
listPackages verbosity cli (Just (Id pkgid)) Nothing
Just m -> listPackages verbosity cli (Just (Substring pkgid_str m)) Nothing
["dot"] -> do
showPackageDot verbosity cli
["find-module", moduleName] -> do
let match = maybe (==moduleName) id (substringCheck moduleName)
listPackages verbosity cli Nothing (Just match)
["latest", pkgid_str] -> do
pkgid <- readGlobPkgId pkgid_str
latestPackage verbosity cli pkgid
["describe", pkgid_str] -> do
pkgarg <- case substringCheck pkgid_str of
Nothing -> liftM Id (readGlobPkgId pkgid_str)
Just m -> return (Substring pkgid_str m)
describePackage verbosity cli pkgarg (fromMaybe False mexpand_pkgroot)
["field", pkgid_str, fields] -> do
pkgarg <- case substringCheck pkgid_str of
Nothing -> liftM Id (readGlobPkgId pkgid_str)
Just m -> return (Substring pkgid_str m)
describeField verbosity cli pkgarg
(splitFields fields) (fromMaybe True mexpand_pkgroot)
["check"] -> do
checkConsistency verbosity cli
["dump"] -> do
dumpPackages verbosity cli (fromMaybe False mexpand_pkgroot)
["recache"] -> do
recache verbosity cli
[] -> do
die ("missing command\n" ++ shortUsage prog)
(_cmd:_) -> do
die ("command-line syntax error\n" ++ shortUsage prog)
parseCheck :: ReadP a a -> String -> String -> IO a
parseCheck parser str what =
case [ x | (x,ys) <- readP_to_S parser str, all isSpace ys ] of
[x] -> return x
_ -> die ("cannot parse \'" ++ str ++ "\' as a " ++ what)
readGlobPkgId :: String -> IO PackageIdentifier
readGlobPkgId str = parseCheck parseGlobPackageId str "package identifier"
parseGlobPackageId :: ReadP r PackageIdentifier
parseGlobPackageId =
parse
+++
(do n <- parse
_ <- string "-*"
return (PackageIdentifier{ pkgName = n, pkgVersion = globVersion }))
-- globVersion means "all versions"
globVersion :: Version
globVersion = Version{ versionBranch=[], versionTags=["*"] }
-- -----------------------------------------------------------------------------
-- Package databases
-- Some commands operate on a single database:
-- register, unregister, expose, hide, trust, distrust
-- however these commands also check the union of the available databases
-- in order to check consistency. For example, register will check that
-- dependencies exist before registering a package.
--
-- Some commands operate on multiple databases, with overlapping semantics:
-- list, describe, field
data PackageDB
= PackageDB {
location, locationAbsolute :: !FilePath,
-- We need both possibly-relative and definately-absolute package
-- db locations. This is because the relative location is used as
-- an identifier for the db, so it is important we do not modify it.
-- On the other hand we need the absolute path in a few places
-- particularly in relation to the ${pkgroot} stuff.
packages :: [InstalledPackageInfo]
}
type PackageDBStack = [PackageDB]
-- A stack of package databases. Convention: head is the topmost
-- in the stack.
allPackagesInStack :: PackageDBStack -> [InstalledPackageInfo]
allPackagesInStack = concatMap packages
getPkgDatabases :: Verbosity
-> Bool -- we are modifying, not reading
-> Bool -- read caches, if available
-> Bool -- expand vars, like ${pkgroot} and $topdir
-> [Flag]
-> IO (PackageDBStack,
-- the real package DB stack: [global,user] ++
-- DBs specified on the command line with -f.
Maybe FilePath,
-- which one to modify, if any
PackageDBStack)
-- the package DBs specified on the command
-- line, or [global,user] otherwise. This
-- is used as the list of package DBs for
-- commands that just read the DB, such as 'list'.
getPkgDatabases verbosity modify use_cache expand_vars my_flags = do
-- first we determine the location of the global package config. On Windows,
-- this is found relative to the ghc-pkg.exe binary, whereas on Unix the
-- location is passed to the binary using the --global-package-db flag by the
-- wrapper script.
let err_msg = "missing --global-package-db option, location of global package database unknown\n"
global_conf <-
case [ f | FlagGlobalConfig f <- my_flags ] of
[] -> do mb_dir <- getLibDir
case mb_dir of
Nothing -> die err_msg
Just dir -> do
r <- lookForPackageDBIn dir
case r of
Nothing -> die ("Can't find package database in " ++ dir)
Just path -> return path
fs -> return (last fs)
-- The value of the $topdir variable used in some package descriptions
-- Note that the way we calculate this is slightly different to how it
-- is done in ghc itself. We rely on the convention that the global
-- package db lives in ghc's libdir.
top_dir <- absolutePath (takeDirectory global_conf)
let no_user_db = FlagNoUserDb `elem` my_flags
mb_user_conf <-
if no_user_db then return Nothing else do
r <- lookForPackageDBIn hasteUserDir
case r of
Nothing -> return (Just (pkgUserDir, False))
Just f -> return (Just (f, True))
-- If the user database doesn't exist, and this command isn't a
-- "modify" command, then we won't attempt to create or use it.
let sys_databases
| Just (user_conf,user_exists) <- mb_user_conf,
modify || user_exists = [user_conf, global_conf]
| otherwise = [global_conf]
e_pkg_path <- tryIO (System.Environment.getEnv "GHC_PACKAGE_PATH")
let env_stack =
case e_pkg_path of
Left _ -> sys_databases
Right path
| last cs == "" -> init cs ++ sys_databases
| otherwise -> cs
where cs = parseSearchPath path
-- The "global" database is always the one at the bottom of the stack.
-- This is the database we modify by default.
virt_global_conf = last env_stack
let db_flags = [ f | Just f <- map is_db_flag my_flags ]
where is_db_flag FlagUser
| Just (user_conf, _user_exists) <- mb_user_conf
= Just user_conf
is_db_flag FlagGlobal = Just virt_global_conf
is_db_flag (FlagConfig f) = Just f
is_db_flag _ = Nothing
let flag_db_names | null db_flags = env_stack
| otherwise = reverse (nub db_flags)
-- For a "modify" command, treat all the databases as
-- a stack, where we are modifying the top one, but it
-- can refer to packages in databases further down the
-- stack.
-- -f flags on the command line add to the database
-- stack, unless any of them are present in the stack
-- already.
let final_stack = filter (`notElem` env_stack)
[ f | FlagConfig f <- reverse my_flags ]
++ env_stack
-- the database we actually modify is the one mentioned
-- rightmost on the command-line.
let to_modify
| not modify = Nothing
| null db_flags = Just virt_global_conf
| otherwise = Just (last db_flags)
db_stack <- sequence
[ do db <- readParseDatabase verbosity mb_user_conf use_cache db_path
if expand_vars then return (mungePackageDBPaths top_dir db)
else return db
| db_path <- final_stack ]
let flag_db_stack = [ db | db_name <- flag_db_names,
db <- db_stack, location db == db_name ]
return (db_stack, to_modify, flag_db_stack)
lookForPackageDBIn :: FilePath -> IO (Maybe FilePath)
lookForPackageDBIn dir = do
let path_dir = dir </> "package.conf.d"
exists_dir <- doesDirectoryExist path_dir
if exists_dir then return (Just path_dir) else do
let path_file = dir </> "package.conf"
exists_file <- doesFileExist path_file
if exists_file then return (Just path_file) else return Nothing
readParseDatabase :: Verbosity
-> Maybe (FilePath,Bool)
-> Bool -- use cache
-> FilePath
-> IO PackageDB
readParseDatabase verbosity mb_user_conf use_cache path
-- the user database (only) is allowed to be non-existent
| Just (user_conf,False) <- mb_user_conf, path == user_conf
= mkPackageDB []
| otherwise
= do e <- tryIO $ getDirectoryContents path
case e of
Left _ -> do
pkgs <- parseMultiPackageConf verbosity path
mkPackageDB pkgs
Right fs
| not use_cache -> ignore_cache (const $ return ())
| otherwise -> do
let cache = path </> cachefilename
tdir <- getModificationTime path
e_tcache <- tryIO $ getModificationTime cache
case e_tcache of
Left ex -> do
when (verbosity > Normal) $
warn ("warning: cannot read cache file " ++ cache ++ ": " ++ show ex)
ignore_cache (const $ return ())
Right tcache -> do
let compareTimestampToCache file =
when (verbosity >= Verbose) $ do
tFile <- getModificationTime file
compareTimestampToCache' file tFile
compareTimestampToCache' file tFile = do
let rel = case tcache `compare` tFile of
LT -> " (NEWER than cache)"
GT -> " (older than cache)"
EQ -> " (same as cache)"
warn ("Timestamp " ++ show tFile
++ " for " ++ file ++ rel)
when (verbosity >= Verbose) $ do
warn ("Timestamp " ++ show tcache ++ " for " ++ cache)
compareTimestampToCache' path tdir
if tcache >= tdir
then do
when (verbosity > Normal) $
infoLn ("using cache: " ++ cache)
pkgs <- myReadBinPackageDB cache
let pkgs' = map convertPackageInfoIn pkgs
mkPackageDB pkgs'
else do
when (verbosity >= Normal) $ do
warn ("WARNING: cache is out of date: "
++ cache)
warn "Use 'ghc-pkg recache' to fix."
ignore_cache compareTimestampToCache
where
ignore_cache :: (FilePath -> IO ()) -> IO PackageDB
ignore_cache checkTime = do
let confs = filter (".conf" `isSuffixOf`) fs
doFile f = do checkTime f
parseSingletonPackageConf verbosity f
pkgs <- mapM doFile $ map (path </>) confs
mkPackageDB pkgs
where
mkPackageDB pkgs = do
path_abs <- absolutePath path
return PackageDB {
location = path,
locationAbsolute = path_abs,
packages = pkgs
}
-- read the package.cache file strictly, to work around a problem with
-- bytestring 0.9.0.x (fixed in 0.9.1.x) where the file wasn't closed
-- after it has been completely read, leading to a sharing violation
-- later.
myReadBinPackageDB :: FilePath -> IO [InstalledPackageInfoString]
myReadBinPackageDB filepath = do
h <- openBinaryFile filepath ReadMode
sz <- hFileSize h
b <- B.hGet h (fromIntegral sz)
hClose h
return $ Bin.runGet Bin.get b
parseMultiPackageConf :: Verbosity -> FilePath -> IO [InstalledPackageInfo]
parseMultiPackageConf verbosity file = do
when (verbosity > Normal) $ infoLn ("reading package database: " ++ file)
str <- readUTF8File file
let pkgs = map convertPackageInfoIn $ read str
Exception.evaluate pkgs
`catchError` \e->
die ("error while parsing " ++ file ++ ": " ++ show e)
parseSingletonPackageConf :: Verbosity -> FilePath -> IO InstalledPackageInfo
parseSingletonPackageConf verbosity file = do
when (verbosity > Normal) $ infoLn ("reading package config: " ++ file)
readUTF8File file >>= fmap fst . parsePackageInfo
cachefilename :: FilePath
cachefilename = "package.cache"
mungePackageDBPaths :: FilePath -> PackageDB -> PackageDB
mungePackageDBPaths top_dir db@PackageDB { packages = pkgs } =
db { packages = map (mungePackagePaths top_dir pkgroot) pkgs }
where
pkgroot = takeDirectory (locationAbsolute db)
-- It so happens that for both styles of package db ("package.conf"
-- files and "package.conf.d" dirs) the pkgroot is the parent directory
-- ${pkgroot}/package.conf or ${pkgroot}/package.conf.d/
-- TODO: This code is duplicated in compiler/main/Packages.lhs
mungePackagePaths :: FilePath -> FilePath
-> InstalledPackageInfo -> InstalledPackageInfo
-- Perform path/URL variable substitution as per the Cabal ${pkgroot} spec
-- (http://www.haskell.org/pipermail/libraries/2009-May/011772.html)
-- Paths/URLs can be relative to ${pkgroot} or ${pkgrooturl}.
-- The "pkgroot" is the directory containing the package database.
--
-- Also perform a similar substitution for the older GHC-specific
-- "$topdir" variable. The "topdir" is the location of the ghc
-- installation (obtained from the -B option).
mungePackagePaths top_dir pkgroot pkg =
pkg {
importDirs = munge_paths (importDirs pkg),
includeDirs = munge_paths (includeDirs pkg),
libraryDirs = munge_paths (libraryDirs pkg),
frameworkDirs = munge_paths (frameworkDirs pkg),
haddockInterfaces = munge_paths (haddockInterfaces pkg),
-- haddock-html is allowed to be either a URL or a file
haddockHTMLs = munge_paths (munge_urls (haddockHTMLs pkg))
}
where
munge_paths = map munge_path
munge_urls = map munge_url
munge_path p
| Just p' <- stripVarPrefix "${pkgroot}" p = pkgroot ++ p'
| Just p' <- stripVarPrefix "$topdir" p = top_dir ++ p'
| otherwise = p
munge_url p
| Just p' <- stripVarPrefix "${pkgrooturl}" p = toUrlPath pkgroot p'
| Just p' <- stripVarPrefix "$httptopdir" p = toUrlPath top_dir p'
| otherwise = p
toUrlPath r p = "file:///"
-- URLs always use posix style '/' separators:
++ FilePath.Posix.joinPath
(r : -- We need to drop a leading "/" or "\\"
-- if there is one:
dropWhile (all isPathSeparator)
(FilePath.splitDirectories p))
-- We could drop the separator here, and then use </> above. However,
-- by leaving it in and using ++ we keep the same path separator
-- rather than letting FilePath change it to use \ as the separator
stripVarPrefix var path = case stripPrefix var path of
Just [] -> Just []
Just cs@(c : _) | isPathSeparator c -> Just cs
_ -> Nothing
-- -----------------------------------------------------------------------------
-- Creating a new package DB
initPackageDB :: FilePath -> Verbosity -> [Flag] -> IO ()
initPackageDB filename verbosity _flags = do
let eexist = die ("cannot create: " ++ filename ++ " already exists")
b1 <- doesFileExist filename
when b1 eexist
b2 <- doesDirectoryExist filename
when b2 eexist
filename_abs <- absolutePath filename
changeDB verbosity [] PackageDB {
location = filename, locationAbsolute = filename_abs,
packages = []
}
-- -----------------------------------------------------------------------------
-- Registering
registerPackage :: FilePath
-> Verbosity
-> [Flag]
-> Bool -- auto_ghci_libs
-> Bool -- expand_env_vars
-> Bool -- update
-> Force
-> IO ()
registerPackage input verbosity my_flags auto_ghci_libs expand_env_vars update force = do
(db_stack, Just to_modify, _flag_dbs) <-
getPkgDatabases verbosity True True False{-expand vars-} my_flags
let
db_to_operate_on = my_head "register" $
filter ((== to_modify).location) db_stack
--
when (auto_ghci_libs && verbosity >= Silent) $
warn "Warning: --auto-ghci-libs is deprecated and will be removed in GHC 7.4"
--
s <-
case input of
"-" -> do
when (verbosity >= Normal) $
info "Reading package info from stdin ... "
-- fix the encoding to UTF-8, since this is an interchange format
hSetEncoding stdin utf8
getContents
f -> do
when (verbosity >= Normal) $
info ("Reading package info from " ++ show f ++ " ... ")
readUTF8File f
expanded <- if expand_env_vars then expandEnvVars s force
else return s
(pkg, ws) <- parsePackageInfo expanded
when (verbosity >= Normal) $
infoLn "done."
-- report any warnings from the parse phase
_ <- reportValidateErrors [] ws
(display (sourcePackageId pkg) ++ ": Warning: ") Nothing
-- validate the expanded pkg, but register the unexpanded
pkgroot <- absolutePath (takeDirectory to_modify)
let top_dir = takeDirectory (location (last db_stack))
pkg_expanded = mungePackagePaths top_dir pkgroot pkg
let truncated_stack = dropWhile ((/= to_modify).location) db_stack
-- truncate the stack for validation, because we don't allow
-- packages lower in the stack to refer to those higher up.
validatePackageConfig pkg_expanded verbosity truncated_stack auto_ghci_libs update force
let
removes = [ RemovePackage p
| p <- packages db_to_operate_on,
sourcePackageId p == sourcePackageId pkg ]
--
changeDB verbosity (removes ++ [AddPackage pkg]) db_to_operate_on
parsePackageInfo
:: String
-> IO (InstalledPackageInfo, [ValidateWarning])
parsePackageInfo str =
case parseInstalledPackageInfo str of
ParseOk warnings ok -> return (ok, ws)
where
ws = [ msg | PWarning msg <- warnings
, not ("Unrecognized field pkgroot" `isPrefixOf` msg) ]
ParseFailed err -> case locatedErrorMsg err of
(Nothing, s) -> die s
(Just l, s) -> die (show l ++ ": " ++ s)
-- -----------------------------------------------------------------------------
-- Making changes to a package database
data DBOp = RemovePackage InstalledPackageInfo
| AddPackage InstalledPackageInfo
| ModifyPackage InstalledPackageInfo
changeDB :: Verbosity -> [DBOp] -> PackageDB -> IO ()
changeDB verbosity cmds db = do
let db' = updateInternalDB db cmds
isfile <- doesFileExist (location db)
if isfile
then writeNewConfig verbosity (location db') (packages db')
else do
createDirectoryIfMissing True (location db)
changeDBDir verbosity cmds db'
updateInternalDB :: PackageDB -> [DBOp] -> PackageDB
updateInternalDB db cmds = db{ packages = foldl do_cmd (packages db) cmds }
where
do_cmd pkgs (RemovePackage p) =
filter ((/= installedPackageId p) . installedPackageId) pkgs
do_cmd pkgs (AddPackage p) = p : pkgs
do_cmd pkgs (ModifyPackage p) =
do_cmd (do_cmd pkgs (RemovePackage p)) (AddPackage p)
changeDBDir :: Verbosity -> [DBOp] -> PackageDB -> IO ()
changeDBDir verbosity cmds db = do
mapM_ do_cmd cmds
updateDBCache verbosity db
where
do_cmd (RemovePackage p) = do
let file = location db </> display (installedPackageId p) <.> "conf"
when (verbosity > Normal) $ infoLn ("removing " ++ file)
removeFileSafe file
do_cmd (AddPackage p) = do
let file = location db </> display (installedPackageId p) <.> "conf"
when (verbosity > Normal) $ infoLn ("writing " ++ file)
writeFileUtf8Atomic file (showInstalledPackageInfo p)
do_cmd (ModifyPackage p) =
do_cmd (AddPackage p)
updateDBCache :: Verbosity -> PackageDB -> IO ()
updateDBCache verbosity db = do
let filename = location db </> cachefilename
when (verbosity > Normal) $
infoLn ("writing cache " ++ filename)
writeBinaryFileAtomic filename (map convertPackageInfoOut (packages db))
`catchIO` \e ->
if isPermissionError e
then die (filename ++ ": you don't have permission to modify this file")
else ioError e
#ifndef mingw32_HOST_OS
status <- getFileStatus filename
setFileTimes (location db) (accessTime status) (modificationTime status)
#endif
-- -----------------------------------------------------------------------------
-- Exposing, Hiding, Trusting, Distrusting, Unregistering are all similar
exposePackage :: PackageIdentifier -> Verbosity -> [Flag] -> Force -> IO ()
exposePackage = modifyPackage (\p -> ModifyPackage p{exposed=True})
hidePackage :: PackageIdentifier -> Verbosity -> [Flag] -> Force -> IO ()
hidePackage = modifyPackage (\p -> ModifyPackage p{exposed=False})
trustPackage :: PackageIdentifier -> Verbosity -> [Flag] -> Force -> IO ()
trustPackage = modifyPackage (\p -> ModifyPackage p{trusted=True})
distrustPackage :: PackageIdentifier -> Verbosity -> [Flag] -> Force -> IO ()
distrustPackage = modifyPackage (\p -> ModifyPackage p{trusted=False})
unregisterPackage :: PackageIdentifier -> Verbosity -> [Flag] -> Force -> IO ()
unregisterPackage = modifyPackage RemovePackage
modifyPackage
:: (InstalledPackageInfo -> DBOp)
-> PackageIdentifier
-> Verbosity
-> [Flag]
-> Force
-> IO ()
modifyPackage fn pkgid verbosity my_flags force = do
(db_stack, Just _to_modify, _flag_dbs) <-
getPkgDatabases verbosity True{-modify-} True{-use cache-} False{-expand vars-} my_flags
(db, ps) <- fmap head $ findPackagesByDB db_stack (Id pkgid)
let
db_name = location db
pkgs = packages db
pids = map sourcePackageId ps
cmds = [ fn pkg | pkg <- pkgs, sourcePackageId pkg `elem` pids ]
new_db = updateInternalDB db cmds
old_broken = brokenPackages (allPackagesInStack db_stack)
rest_of_stack = filter ((/= db_name) . location) db_stack
new_stack = new_db : rest_of_stack
new_broken = map sourcePackageId (brokenPackages (allPackagesInStack new_stack))
newly_broken = filter (`notElem` map sourcePackageId old_broken) new_broken
--
when (not (null newly_broken)) $
dieOrForceAll force ("unregistering " ++ display pkgid ++
" would break the following packages: "
++ unwords (map display newly_broken))
changeDB verbosity cmds db
recache :: Verbosity -> [Flag] -> IO ()
recache verbosity my_flags = do
(db_stack, Just to_modify, _flag_dbs) <-
getPkgDatabases verbosity True{-modify-} False{-no cache-} False{-expand vars-} my_flags
let
db_to_operate_on = my_head "recache" $
filter ((== to_modify).location) db_stack
--
changeDB verbosity [] db_to_operate_on
-- -----------------------------------------------------------------------------
-- Listing packages
listPackages :: Verbosity -> [Flag] -> Maybe PackageArg
-> Maybe (String->Bool)
-> IO ()
listPackages verbosity my_flags mPackageName mModuleName = do
let simple_output = FlagSimpleOutput `elem` my_flags
(db_stack, _, flag_db_stack) <-
getPkgDatabases verbosity False True{-use cache-} False{-expand vars-} my_flags
let db_stack_filtered -- if a package is given, filter out all other packages
| Just this <- mPackageName =
[ db{ packages = filter (this `matchesPkg`) (packages db) }
| db <- flag_db_stack ]
| Just match <- mModuleName = -- packages which expose mModuleName
[ db{ packages = filter (match `exposedInPkg`) (packages db) }
| db <- flag_db_stack ]
| otherwise = flag_db_stack
db_stack_sorted
= [ db{ packages = sort_pkgs (packages db) }
| db <- db_stack_filtered ]
where sort_pkgs = sortBy cmpPkgIds
cmpPkgIds pkg1 pkg2 =
case pkgName p1 `compare` pkgName p2 of
LT -> LT
GT -> GT
EQ -> pkgVersion p1 `compare` pkgVersion p2
where (p1,p2) = (sourcePackageId pkg1, sourcePackageId pkg2)
stack = reverse db_stack_sorted
match `exposedInPkg` pkg = any match (map display $ exposedModules pkg)
pkg_map = allPackagesInStack db_stack
broken = map sourcePackageId (brokenPackages pkg_map)
show_normal PackageDB{ location = db_name, packages = pkg_confs } =
do hPutStrLn stdout (db_name ++ ":")
if null pp_pkgs
then hPutStrLn stdout " (no packages)"
else hPutStrLn stdout $ unlines (map (" " ++) pp_pkgs)
where
-- Sort using instance Ord PackageId
pp_pkgs = map pp_pkg . sortBy (comparing installedPackageId) $ pkg_confs
pp_pkg p
| sourcePackageId p `elem` broken = printf "{%s}" doc
| exposed p = doc
| otherwise = printf "(%s)" doc
where doc | verbosity >= Verbose = printf "%s (%s)" pkg ipid
| otherwise = pkg
where
InstalledPackageId ipid = installedPackageId p
pkg = display (sourcePackageId p)
show_simple = simplePackageList my_flags . allPackagesInStack
when (not (null broken) && not simple_output && verbosity /= Silent) $ do
prog <- getProgramName
warn ("WARNING: there are broken packages. Run '" ++ prog ++ " check' for more details.")
if simple_output then show_simple stack else do
#if defined(mingw32_HOST_OS) || defined(BOOTSTRAPPING)
mapM_ show_normal stack
#else
let
show_colour withF db =
mconcat $ map (<#> termText "\n") $
(termText (location db) :
map (termText " " <#>) (map pp_pkg (packages db)))
where
pp_pkg p
| sourcePackageId p `elem` broken = withF Red doc
| exposed p = doc
| otherwise = withF Blue doc
where doc | verbosity >= Verbose
= termText (printf "%s (%s)" pkg ipid)
| otherwise
= termText pkg
where
InstalledPackageId ipid = installedPackageId p
pkg = display (sourcePackageId p)
is_tty <- hIsTerminalDevice stdout
if not is_tty
then mapM_ show_normal stack
else do tty <- Terminfo.setupTermFromEnv
case Terminfo.getCapability tty withForegroundColor of
Nothing -> mapM_ show_normal stack
Just w -> runTermOutput tty $ mconcat $
map (show_colour w) stack
#endif
simplePackageList :: [Flag] -> [InstalledPackageInfo] -> IO ()
simplePackageList my_flags pkgs = do
let showPkg = if FlagNamesOnly `elem` my_flags then display . pkgName
else display
-- Sort using instance Ord PackageId
strs = map showPkg $ sort $ map sourcePackageId pkgs
when (not (null pkgs)) $
hPutStrLn stdout $ concat $ intersperse " " strs
showPackageDot :: Verbosity -> [Flag] -> IO ()
showPackageDot verbosity myflags = do
(_, _, flag_db_stack) <-
getPkgDatabases verbosity False True{-use cache-} False{-expand vars-} myflags
let all_pkgs = allPackagesInStack flag_db_stack
ipix = PackageIndex.fromList all_pkgs
putStrLn "digraph {"
let quote s = '"':s ++ "\""
mapM_ putStrLn [ quote from ++ " -> " ++ quote to
| p <- all_pkgs,
let from = display (sourcePackageId p),
depid <- depends p,
Just dep <- [PackageIndex.lookupInstalledPackageId ipix depid],
let to = display (sourcePackageId dep)
]
putStrLn "}"
-- -----------------------------------------------------------------------------
-- Prints the highest (hidden or exposed) version of a package
latestPackage :: Verbosity -> [Flag] -> PackageIdentifier -> IO ()
latestPackage verbosity my_flags pkgid = do
(_, _, flag_db_stack) <-
getPkgDatabases verbosity False True{-use cache-} False{-expand vars-} my_flags
ps <- findPackages flag_db_stack (Id pkgid)
case ps of
[] -> die "no matches"
_ -> show_pkg . maximum . map sourcePackageId $ ps
where
show_pkg pid = hPutStrLn stdout (display pid)
-- -----------------------------------------------------------------------------
-- Describe
describePackage :: Verbosity -> [Flag] -> PackageArg -> Bool -> IO ()
describePackage verbosity my_flags pkgarg expand_pkgroot = do
(_, _, flag_db_stack) <-
getPkgDatabases verbosity False True{-use cache-} expand_pkgroot my_flags
dbs <- findPackagesByDB flag_db_stack pkgarg
doDump expand_pkgroot [ (pkg, locationAbsolute db)
| (db, pkgs) <- dbs, pkg <- pkgs ]
dumpPackages :: Verbosity -> [Flag] -> Bool -> IO ()
dumpPackages verbosity my_flags expand_pkgroot = do
(_, _, flag_db_stack) <-
getPkgDatabases verbosity False True{-use cache-} expand_pkgroot my_flags
doDump expand_pkgroot [ (pkg, locationAbsolute db)
| db <- flag_db_stack, pkg <- packages db ]
doDump :: Bool -> [(InstalledPackageInfo, FilePath)] -> IO ()
doDump expand_pkgroot pkgs = do
-- fix the encoding to UTF-8, since this is an interchange format
hSetEncoding stdout utf8
putStrLn $
intercalate "---\n"
[ if expand_pkgroot
then showInstalledPackageInfo pkg
else showInstalledPackageInfo pkg ++ pkgrootField
| (pkg, pkgloc) <- pkgs
, let pkgroot = takeDirectory pkgloc
pkgrootField = "pkgroot: " ++ show pkgroot ++ "\n" ]
-- PackageId is can have globVersion for the version
findPackages :: PackageDBStack -> PackageArg -> IO [InstalledPackageInfo]
findPackages db_stack pkgarg
= fmap (concatMap snd) $ findPackagesByDB db_stack pkgarg
findPackagesByDB :: PackageDBStack -> PackageArg
-> IO [(PackageDB, [InstalledPackageInfo])]
findPackagesByDB db_stack pkgarg
= case [ (db, matched)
| db <- db_stack,
let matched = filter (pkgarg `matchesPkg`) (packages db),
not (null matched) ] of
[] -> die ("cannot find package " ++ pkg_msg pkgarg)
ps -> return ps
where
pkg_msg (Id pkgid) = display pkgid
pkg_msg (Substring pkgpat _) = "matching " ++ pkgpat
matches :: PackageIdentifier -> PackageIdentifier -> Bool
pid `matches` pid'
= (pkgName pid == pkgName pid')
&& (pkgVersion pid == pkgVersion pid' || not (realVersion pid))
realVersion :: PackageIdentifier -> Bool
realVersion pkgid = versionBranch (pkgVersion pkgid) /= []
-- when versionBranch == [], this is a glob
matchesPkg :: PackageArg -> InstalledPackageInfo -> Bool
(Id pid) `matchesPkg` pkg = pid `matches` sourcePackageId pkg
(Substring _ m) `matchesPkg` pkg = m (display (sourcePackageId pkg))
-- -----------------------------------------------------------------------------
-- Field
describeField :: Verbosity -> [Flag] -> PackageArg -> [String] -> Bool -> IO ()
describeField verbosity my_flags pkgarg fields expand_pkgroot = do
(_, _, flag_db_stack) <-
getPkgDatabases verbosity False True{-use cache-} expand_pkgroot my_flags
fns <- mapM toField fields
ps <- findPackages flag_db_stack pkgarg
mapM_ (selectFields fns) ps
where showFun = if FlagSimpleOutput `elem` my_flags
then showSimpleInstalledPackageInfoField
else showInstalledPackageInfoField
toField f = case showFun f of
Nothing -> die ("unknown field: " ++ f)
Just fn -> return fn
selectFields fns pinfo = mapM_ (\fn->putStrLn (fn pinfo)) fns
-- -----------------------------------------------------------------------------
-- Check: Check consistency of installed packages
checkConsistency :: Verbosity -> [Flag] -> IO ()
checkConsistency verbosity my_flags = do
(db_stack, _, _) <-
getPkgDatabases verbosity True True{-use cache-} True{-expand vars-} my_flags
-- check behaves like modify for the purposes of deciding which
-- databases to use, because ordering is important.
let simple_output = FlagSimpleOutput `elem` my_flags
let pkgs = allPackagesInStack db_stack
checkPackage p = do
(_,es,ws) <- runValidate $ checkPackageConfig p verbosity db_stack False True
if null es
then do when (not simple_output) $ do
_ <- reportValidateErrors [] ws "" Nothing
return ()
return []
else do
when (not simple_output) $ do
reportError ("There are problems in package " ++ display (sourcePackageId p) ++ ":")
_ <- reportValidateErrors es ws " " Nothing
return ()
return [p]
broken_pkgs <- concat `fmap` mapM checkPackage pkgs
let filterOut pkgs1 pkgs2 = filter not_in pkgs2
where not_in p = sourcePackageId p `notElem` all_ps
all_ps = map sourcePackageId pkgs1
let not_broken_pkgs = filterOut broken_pkgs pkgs
(_, trans_broken_pkgs) = closure [] not_broken_pkgs
all_broken_pkgs = broken_pkgs ++ trans_broken_pkgs
when (not (null all_broken_pkgs)) $ do
if simple_output
then simplePackageList my_flags all_broken_pkgs
else do
reportError ("\nThe following packages are broken, either because they have a problem\n"++
"listed above, or because they depend on a broken package.")
mapM_ (hPutStrLn stderr . display . sourcePackageId) all_broken_pkgs
when (not (null all_broken_pkgs)) $ exitWith (ExitFailure 1)
closure :: [InstalledPackageInfo] -> [InstalledPackageInfo]
-> ([InstalledPackageInfo], [InstalledPackageInfo])
closure pkgs db_stack = go pkgs db_stack
where
go avail not_avail =
case partition (depsAvailable avail) not_avail of
([], not_avail') -> (avail, not_avail')
(new_avail, not_avail') -> go (new_avail ++ avail) not_avail'
depsAvailable :: [InstalledPackageInfo] -> InstalledPackageInfo
-> Bool
depsAvailable pkgs_ok pkg = null dangling
where dangling = filter (`notElem` pids) (depends pkg)
pids = map installedPackageId pkgs_ok
-- we want mutually recursive groups of package to show up
-- as broken. (#1750)
brokenPackages :: [InstalledPackageInfo] -> [InstalledPackageInfo]
brokenPackages pkgs = snd (closure [] pkgs)
-- -----------------------------------------------------------------------------
-- Manipulating package.conf files
type InstalledPackageInfoString = InstalledPackageInfo_ String
convertPackageInfoOut :: InstalledPackageInfo -> InstalledPackageInfoString
convertPackageInfoOut
(pkgconf@(InstalledPackageInfo { exposedModules = e,
hiddenModules = h })) =
pkgconf{ exposedModules = map display e,
hiddenModules = map display h }
convertPackageInfoIn :: InstalledPackageInfoString -> InstalledPackageInfo
convertPackageInfoIn
(pkgconf@(InstalledPackageInfo { exposedModules = e,
hiddenModules = h })) =
pkgconf{ exposedModules = map convert e,
hiddenModules = map convert h }
where convert = fromJust . simpleParse
writeNewConfig :: Verbosity -> FilePath -> [InstalledPackageInfo] -> IO ()
writeNewConfig verbosity filename ipis = do
when (verbosity >= Normal) $
info "Writing new package config file... "
createDirectoryIfMissing True $ takeDirectory filename
let shown = concat $ intersperse ",\n "
$ map (show . convertPackageInfoOut) ipis
fileContents = "[" ++ shown ++ "\n]"
writeFileUtf8Atomic filename fileContents
`catchIO` \e ->
if isPermissionError e
then die (filename ++ ": you don't have permission to modify this file")
else ioError e
when (verbosity >= Normal) $
infoLn "done."
-----------------------------------------------------------------------------
-- Sanity-check a new package config, and automatically build GHCi libs
-- if requested.
type ValidateError = (Force,String)
type ValidateWarning = String
newtype Validate a = V { runValidate :: IO (a, [ValidateError],[ValidateWarning]) }
instance Functor Validate where
fmap = liftM
instance Applicative Validate where
pure = return
(<*>) = ap
instance Monad Validate where
return a = V $ return (a, [], [])
m >>= k = V $ do
(a, es, ws) <- runValidate m
(b, es', ws') <- runValidate (k a)
return (b,es++es',ws++ws')
verror :: Force -> String -> Validate ()
verror f s = V (return ((),[(f,s)],[]))
vwarn :: String -> Validate ()
vwarn s = V (return ((),[],["Warning: " ++ s]))
liftIO :: IO a -> Validate a
liftIO k = V (k >>= \a -> return (a,[],[]))
-- returns False if we should die
reportValidateErrors :: [ValidateError] -> [ValidateWarning]
-> String -> Maybe Force -> IO Bool
reportValidateErrors es ws prefix mb_force = do
mapM_ (warn . (prefix++)) ws
oks <- mapM report es
return (and oks)
where
report (f,s)
| Just force <- mb_force
= if (force >= f)
then do reportError (prefix ++ s ++ " (ignoring)")
return True
else if f < CannotForce
then do reportError (prefix ++ s ++ " (use --force to override)")
return False
else do reportError err
return False
| otherwise = do reportError err
return False
where
err = prefix ++ s
validatePackageConfig :: InstalledPackageInfo
-> Verbosity
-> PackageDBStack
-> Bool -- auto-ghc-libs
-> Bool -- update, or check
-> Force
-> IO ()
validatePackageConfig pkg verbosity db_stack auto_ghci_libs update force = do
(_,es,ws) <- runValidate $ checkPackageConfig pkg verbosity db_stack auto_ghci_libs update
ok <- reportValidateErrors es ws (display (sourcePackageId pkg) ++ ": ") (Just force)
when (not ok) $ exitWith (ExitFailure 1)
checkPackageConfig :: InstalledPackageInfo
-> Verbosity
-> PackageDBStack
-> Bool -- auto-ghc-libs
-> Bool -- update, or check
-> Validate ()
checkPackageConfig pkg verbosity db_stack auto_ghci_libs update = do
checkInstalledPackageId pkg db_stack update
checkPackageId pkg
checkDuplicates db_stack pkg update
mapM_ (checkDep db_stack) (depends pkg)
checkDuplicateDepends (depends pkg)
mapM_ (checkDir False "import-dirs") (importDirs pkg)
mapM_ (checkDir True "library-dirs") (libraryDirs pkg)
mapM_ (checkDir True "include-dirs") (includeDirs pkg)
mapM_ (checkDir True "framework-dirs") (frameworkDirs pkg)
mapM_ (checkFile True "haddock-interfaces") (haddockInterfaces pkg)
mapM_ (checkDirURL True "haddock-html") (haddockHTMLs pkg)
checkModules pkg
mapM_ (checkHSLib verbosity (libraryDirs pkg) auto_ghci_libs) (hsLibraries pkg)
-- ToDo: check these somehow?
-- extra_libraries :: [String],
-- c_includes :: [String],
checkInstalledPackageId :: InstalledPackageInfo -> PackageDBStack -> Bool
-> Validate ()
checkInstalledPackageId ipi db_stack update = do
let ipid@(InstalledPackageId str) = installedPackageId ipi
when (null str) $ verror CannotForce "missing id field"
let dups = [ p | p <- allPackagesInStack db_stack,
installedPackageId p == ipid ]
when (not update && not (null dups)) $
verror CannotForce $
"package(s) with this id already exist: " ++
unwords (map (display.packageId) dups)
-- When the package name and version are put together, sometimes we can
-- end up with a package id that cannot be parsed. This will lead to
-- difficulties when the user wants to refer to the package later, so
-- we check that the package id can be parsed properly here.
checkPackageId :: InstalledPackageInfo -> Validate ()
checkPackageId ipi =
let str = display (sourcePackageId ipi) in
case [ x :: PackageIdentifier | (x,ys) <- readP_to_S parse str, all isSpace ys ] of
[_] -> return ()
[] -> verror CannotForce ("invalid package identifier: " ++ str)
_ -> verror CannotForce ("ambiguous package identifier: " ++ str)
checkDuplicates :: PackageDBStack -> InstalledPackageInfo -> Bool -> Validate ()
checkDuplicates db_stack pkg update = do
let
pkgid = sourcePackageId pkg
pkgs = packages (head db_stack)
--
-- Check whether this package id already exists in this DB
--
when (not update && (pkgid `elem` map sourcePackageId pkgs)) $
verror CannotForce $
"package " ++ display pkgid ++ " is already installed"
let
uncasep = map toLower . display
dups = filter ((== uncasep pkgid) . uncasep) (map sourcePackageId pkgs)
when (not update && not (null dups)) $ verror ForceAll $
"Package names may be treated case-insensitively in the future.\n"++
"Package " ++ display pkgid ++
" overlaps with: " ++ unwords (map display dups)
checkDir, checkFile, checkDirURL :: Bool -> String -> FilePath -> Validate ()
checkDir = checkPath False True
checkFile = checkPath False False
checkDirURL = checkPath True True
checkPath :: Bool -> Bool -> Bool -> String -> FilePath -> Validate ()
checkPath url_ok is_dir warn_only thisfield d
| url_ok && ("http://" `isPrefixOf` d
|| "https://" `isPrefixOf` d) = return ()
| url_ok
, Just d' <- stripPrefix "file://" d
= checkPath False is_dir warn_only thisfield d'
-- Note: we don't check for $topdir/${pkgroot} here. We rely on these
-- variables having been expanded already, see mungePackagePaths.
| isRelative d = verror ForceFiles $
thisfield ++ ": " ++ d ++ " is a relative path which "
++ "makes no sense (as there is nothing for it to be "
++ "relative to). You can make paths relative to the "
++ "package database itself by using ${pkgroot}."
-- relative paths don't make any sense; #4134
| otherwise = do
there <- liftIO $ if is_dir then doesDirectoryExist d else doesFileExist d
when (not there) $
let msg = thisfield ++ ": " ++ d ++ " doesn't exist or isn't a "
++ if is_dir then "directory" else "file"
in
if warn_only
then vwarn msg
else verror ForceFiles msg
checkDep :: PackageDBStack -> InstalledPackageId -> Validate ()
checkDep db_stack pkgid
| pkgid `elem` pkgids = return ()
| otherwise = verror ForceAll ("dependency \"" ++ display pkgid
++ "\" doesn't exist")
where
all_pkgs = allPackagesInStack db_stack
pkgids = map installedPackageId all_pkgs
checkDuplicateDepends :: [InstalledPackageId] -> Validate ()
checkDuplicateDepends deps
| null dups = return ()
| otherwise = verror ForceAll ("package has duplicate dependencies: " ++
unwords (map display dups))
where
dups = [ p | (p:_:_) <- group (sort deps) ]
checkHSLib :: Verbosity -> [String] -> Bool -> String -> Validate ()
checkHSLib verbosity dirs auto_ghci_libs lib = do
let batch_lib_file = "lib" ++ lib ++ ".a"
filenames = ["lib" ++ lib ++ ".a",
"lib" ++ lib ++ ".p_a",
"lib" ++ lib ++ "-ghc" ++ showVersion ghcVersion ++ ".so",
"lib" ++ lib ++ "-ghc" ++ showVersion ghcVersion ++ ".dylib",
lib ++ "-ghc" ++ showVersion ghcVersion ++ ".dll"]
m <- liftIO $ doesFileExistOnPath filenames dirs
case m of
Nothing -> verror ForceFiles ("cannot find any of " ++ show filenames ++
" on library path")
Just dir -> liftIO $ checkGHCiLib verbosity dir batch_lib_file lib auto_ghci_libs
doesFileExistOnPath :: [FilePath] -> [FilePath] -> IO (Maybe FilePath)
doesFileExistOnPath filenames paths = go fullFilenames
where fullFilenames = [ (path, path </> filename)
| filename <- filenames
, path <- paths ]
go [] = return Nothing
go ((p, fp) : xs) = do b <- doesFileExist fp
if b then return (Just p) else go xs
checkModules :: InstalledPackageInfo -> Validate ()
checkModules pkg = do
mapM_ findModule (exposedModules pkg ++ hiddenModules pkg)
where
findModule modl =
-- there's no interface file for GHC.Prim
unless (modl == fromString "GHC.Prim") $ do
let files = [ toFilePath modl <.> extension
| extension <- ["hi", "p_hi", "dyn_hi" ] ]
m <- liftIO $ doesFileExistOnPath files (importDirs pkg)
when (isNothing m) $
verror ForceFiles ("cannot find any of " ++ show files)
checkGHCiLib :: Verbosity -> String -> String -> String -> Bool -> IO ()
checkGHCiLib verbosity batch_lib_dir batch_lib_file lib auto_build
| auto_build = autoBuildGHCiLib verbosity batch_lib_dir batch_lib_file ghci_lib_file
| otherwise = return ()
where
ghci_lib_file = lib <.> "o"
-- automatically build the GHCi version of a batch lib,
-- using ld --whole-archive.
autoBuildGHCiLib :: Verbosity -> String -> String -> String -> IO ()
autoBuildGHCiLib verbosity dir batch_file ghci_file = do
let ghci_lib_file = dir ++ '/':ghci_file
batch_lib_file = dir ++ '/':batch_file
when (verbosity >= Normal) $
info ("building GHCi library " ++ ghci_lib_file ++ "...")
#if defined(darwin_HOST_OS)
r <- rawSystem "ld" ["-r","-x","-o",ghci_lib_file,"-all_load",batch_lib_file]
#elif defined(mingw32_HOST_OS)
execDir <- getLibDir
r <- rawSystem (maybe "" (++"/gcc-lib/") execDir++"ld") ["-r","-x","-o",ghci_lib_file,"--whole-archive",batch_lib_file]
#else
r <- rawSystem "ld" ["-r","-x","-o",ghci_lib_file,"--whole-archive",batch_lib_file]
#endif
when (r /= ExitSuccess) $ exitWith r
when (verbosity >= Normal) $
infoLn (" done.")
-- -----------------------------------------------------------------------------
-- Searching for modules
#if not_yet
findModules :: [FilePath] -> IO [String]
findModules paths =
mms <- mapM searchDir paths
return (concat mms)
searchDir path prefix = do
fs <- getDirectoryEntries path `catchIO` \_ -> return []
searchEntries path prefix fs
searchEntries path prefix [] = return []
searchEntries path prefix (f:fs)
| looks_like_a_module = do
ms <- searchEntries path prefix fs
return (prefix `joinModule` f : ms)
| looks_like_a_component = do
ms <- searchDir (path </> f) (prefix `joinModule` f)
ms' <- searchEntries path prefix fs
return (ms ++ ms')
| otherwise
searchEntries path prefix fs
where
(base,suffix) = splitFileExt f
looks_like_a_module =
suffix `elem` haskell_suffixes &&
all okInModuleName base
looks_like_a_component =
null suffix && all okInModuleName base
okInModuleName c
#endif
-- ---------------------------------------------------------------------------
-- expanding environment variables in the package configuration
expandEnvVars :: String -> Force -> IO String
expandEnvVars str0 force = go str0 ""
where
go "" acc = return $! reverse acc
go ('$':'{':str) acc | (var, '}':rest) <- break close str
= do value <- lookupEnvVar var
go rest (reverse value ++ acc)
where close c = c == '}' || c == '\n' -- don't span newlines
go (c:str) acc
= go str (c:acc)
lookupEnvVar :: String -> IO String
lookupEnvVar "pkgroot" = return "${pkgroot}" -- these two are special,
lookupEnvVar "pkgrooturl" = return "${pkgrooturl}" -- we don't expand them
lookupEnvVar nm =
catchIO (System.Environment.getEnv nm)
(\ _ -> do dieOrForceAll force ("Unable to expand variable " ++
show nm)
return "")
-----------------------------------------------------------------------------
getProgramName :: IO String
getProgramName = liftM (`withoutSuffix` ".bin") getProgName
where str `withoutSuffix` suff
| suff `isSuffixOf` str = take (length str - length suff) str
| otherwise = str
bye :: String -> IO a
bye s = putStr s >> exitWith ExitSuccess
die :: String -> IO a
die = dieWith 1
dieWith :: Int -> String -> IO a
dieWith ec s = do
prog <- getProgramName
reportError (prog ++ ": " ++ s)
exitWith (ExitFailure ec)
dieOrForceAll :: Force -> String -> IO ()
dieOrForceAll ForceAll s = ignoreError s
dieOrForceAll _other s = dieForcible s
warn :: String -> IO ()
warn = reportError
-- send info messages to stdout
infoLn :: String -> IO ()
infoLn = putStrLn
info :: String -> IO ()
info = putStr
ignoreError :: String -> IO ()
ignoreError s = reportError (s ++ " (ignoring)")
reportError :: String -> IO ()
reportError s = do hFlush stdout; hPutStrLn stderr s
dieForcible :: String -> IO ()
dieForcible s = die (s ++ " (use --force to override)")
my_head :: String -> [a] -> a
my_head s [] = error s
my_head _ (x : _) = x
getLibDir :: IO (Maybe String)
getLibDir = return $ Just hasteSysDir
-----------------------------------------
-- Adapted from ghc/compiler/utils/Panic
installSignalHandlers :: IO ()
installSignalHandlers = do
threadid <- myThreadId
let
interrupt = Exception.throwTo threadid
(Exception.ErrorCall "interrupted")
--
#if !defined(mingw32_HOST_OS)
_ <- installHandler sigQUIT (Catch interrupt) Nothing
_ <- installHandler sigINT (Catch interrupt) Nothing
return ()
#else
-- GHC 6.3+ has support for console events on Windows
-- NOTE: running GHCi under a bash shell for some reason requires
-- you to press Ctrl-Break rather than Ctrl-C to provoke
-- an interrupt. Ctrl-C is getting blocked somewhere, I don't know
-- why --SDM 17/12/2004
let sig_handler ControlC = interrupt
sig_handler Break = interrupt
sig_handler _ = return ()
_ <- installHandler (Catch sig_handler)
return ()
#endif
#if mingw32_HOST_OS || mingw32_TARGET_OS
throwIOIO :: Exception.IOException -> IO a
throwIOIO = Exception.throwIO
#endif
catchIO :: IO a -> (Exception.IOException -> IO a) -> IO a
catchIO = Exception.catch
catchError :: IO a -> (String -> IO a) -> IO a
catchError io handler = io `Exception.catch` handler'
where handler' (Exception.ErrorCall err) = handler err
tryIO :: IO a -> IO (Either Exception.IOException a)
tryIO = Exception.try
writeBinaryFileAtomic :: Bin.Binary a => FilePath -> a -> IO ()
writeBinaryFileAtomic targetFile obj =
withFileAtomic targetFile $ \h -> do
hSetBinaryMode h True
B.hPutStr h (Bin.encode obj)
writeFileUtf8Atomic :: FilePath -> String -> IO ()
writeFileUtf8Atomic targetFile content =
withFileAtomic targetFile $ \h -> do
hSetEncoding h utf8
hPutStr h content
-- copied from Cabal's Distribution.Simple.Utils, except that we want
-- to use text files here, rather than binary files.
withFileAtomic :: FilePath -> (Handle -> IO ()) -> IO ()
withFileAtomic targetFile write_content = do
(newFile, newHandle) <- openNewFile targetDir template
do write_content newHandle
hClose newHandle
#if mingw32_HOST_OS || mingw32_TARGET_OS
renameFile newFile targetFile
-- If the targetFile exists then renameFile will fail
`catchIO` \err -> do
exists <- doesFileExist targetFile
if exists
then do removeFileSafe targetFile
-- Big fat hairy race condition
renameFile newFile targetFile
-- If the removeFile succeeds and the renameFile fails
-- then we've lost the atomic property.
else throwIOIO err
#else
renameFile newFile targetFile
#endif
`Exception.onException` do hClose newHandle
removeFileSafe newFile
where
template = targetName <.> "tmp"
targetDir | null targetDir_ = "."
| otherwise = targetDir_
--TODO: remove this when takeDirectory/splitFileName is fixed
-- to always return a valid dir
(targetDir_,targetName) = splitFileName targetFile
openNewFile :: FilePath -> String -> IO (FilePath, Handle)
openNewFile dir template = do
-- this was added to System.IO in 6.12.1
-- we must use this version because the version below opens the file
-- in binary mode.
openTempFileWithDefaultPermissions dir template
-- | The function splits the given string to substrings
-- using 'isSearchPathSeparator'.
parseSearchPath :: String -> [FilePath]
parseSearchPath path = split path
where
split :: String -> [String]
split s =
case rest' of
[] -> [chunk]
_:rest -> chunk : split rest
where
chunk =
case chunk' of
#ifdef mingw32_HOST_OS
('\"':xs@(_:_)) | last xs == '\"' -> init xs
#endif
_ -> chunk'
(chunk', rest') = break isSearchPathSeparator s
readUTF8File :: FilePath -> IO String
readUTF8File file = do
h <- openFile file ReadMode
-- fix the encoding to UTF-8
hSetEncoding h utf8
hGetContents h
-- removeFileSave doesn't throw an exceptions, if the file is already deleted
removeFileSafe :: FilePath -> IO ()
removeFileSafe fn =
removeFile fn `catchIO` \ e ->
when (not $ isDoesNotExistError e) $ ioError e
absolutePath :: FilePath -> IO FilePath
absolutePath path = return . normalise . (</> path) =<< getCurrentDirectory
-- | Only global packages may be marked as relocatable!
-- May break horribly for general use, only reliable for Haste base packages.
relocate :: [String] -> String -> Sh.Shell ()
relocate packages pkg = do
pi <- Sh.run hastePkgBinary (packages ++ ["describe", pkg]) ""
Sh.run_ hastePkgBinary (packages ++ ["update", "-", "--force", "--global"])
(reloc pi)
where
reloc = unlines . map fixPath . lines
fixPath s
| isKey "library-dirs: " s = prefix s "library-dirs" importDir
| isKey "import-dirs: " s = prefix s "import-dirs" importDir
| isKey "haddock-interfaces: " s = prefix s "haddock-interfaces" importDir
| isKey "haddock-html: " s = prefix s "haddock-html" importDir
| isKey "include-dirs: " s = "include-dirs: " ++ includeDir
| otherwise = s
prefix s pfx path = pfx ++ ": " ++ path </> stripPrefix s
stripPrefix s
| os == "darwin" =
case take 3 $ reverse $ splitPath s of
[third, second, first] -> first </> second </> third
| otherwise =
case take 2 $ reverse $ splitPath s of
[second, first] -> first </> second
isKey _ "" =
False
isKey key str =
and $ zipWith (==) key str
importDir
| os == "linux" = "${pkgroot}" </> "libraries" </> "lib"
| otherwise = "${pkgroot}" </> "libraries"
includeDir = "${pkgroot}" </> "include"
| akru/haste-compiler | utils/haste-pkg/haste-pkg.hs | bsd-3-clause | 71,975 | 3 | 99 | 19,824 | 16,873 | 8,499 | 8,374 | 1,279 | 32 |
--
-- Copyright © 2013-2014 Anchor Systems, Pty Ltd and Others
--
-- The code in this file, and the program it is a part of, is
-- made available to you by its authors as open source software:
-- you can redistribute it and/or modify it under the terms of
-- the 3-clause BSD licence.
--
{-# OPTIONS_GHC -fno-warn-orphans #-}
module Marquise.IO.SpoolFile
( newRandomPointsSpoolFile
, newRandomContentsSpoolFile
) where
import Control.Applicative
import Control.Concurrent (threadDelay)
import Control.Monad.State
import Data.ByteString (ByteString)
import qualified Data.ByteString as S
import qualified Data.ByteString.Lazy as L
import qualified Data.ByteString.Lazy as LB
import Data.Maybe
import Marquise.Classes
import Marquise.IO.FFI
import Marquise.Types
import System.Directory
import System.FilePath.Posix
import System.IO
import System.IO.Unsafe
import System.Posix.Files
import System.Posix.IO (closeFd)
import System.Posix.Temp
import System.Posix.Types (Fd)
instance MarquiseSpoolFileMonad IO where
randomSpoolFiles sn =
SpoolFiles <$> newRandomPointsSpoolFile sn
<*> newRandomContentsSpoolFile sn
createDirectories sn =
mapM_ (createDirectoryIfMissing True . ($sn))
[ newPointsDir
, newContentsDir
, curPointsDir
, curContentsDir]
appendPoints spools = doAppend (pointsSpoolFile spools)
appendContents spools = doAppend (contentsSpoolFile spools)
nextPoints sn = nextSpoolContents (newPointsDir sn) (curPointsDir sn)
nextContents sn = nextSpoolContents (newContentsDir sn) (curContentsDir sn)
close _ = c_sync
newRandomSpoolFile :: FilePath -> IO FilePath
newRandomSpoolFile path = do
(spool_file, handle) <- mkstemp path
hClose handle
return spool_file
-- | Creates and returns a new points spool file from a namespace
newRandomPointsSpoolFile :: SpoolName -> IO FilePath
newRandomPointsSpoolFile = newRandomSpoolFile . newPointsDir
-- | Creates and returns a new contents spool file from a namespace
newRandomContentsSpoolFile :: SpoolName -> IO FilePath
newRandomContentsSpoolFile = newRandomSpoolFile . newContentsDir
-- | Grab the next avaliable spool file, providing that file as a lazy
-- bytestring and an action to close it, wiping the file.
nextSpoolContents :: FilePath -> FilePath -> IO (Maybe (L.ByteString, IO ()))
nextSpoolContents new_dir cur_dir = do
-- First check for any work already in the work spool dir.
work <- tryCurDir cur_dir
case work of
Nothing ->
-- No existing work, get some new work out of the spool
-- directory then.
rotate new_dir cur_dir >> return Nothing
Just (fp, lock_fd) -> do
threadDelay 100000 -- Ensure that any slow writes are done
contents <- LB.readFile fp
let close_f = removeLink fp >> closeFd lock_fd
return $ Just (contents, close_f)
-- | Check the work directory for any outstanding work, if there is a potential
-- candidate, lock it. If that fails, try the next.
tryCurDir :: FilePath -> IO (Maybe (FilePath, Fd))
tryCurDir cur_dir =
listToMaybe . catMaybes <$> (getAbsoluteDirectoryFiles cur_dir
>>= mapM lazyLock)
where
lazyLock :: FilePath -> IO (Maybe (FilePath, Fd))
lazyLock fp = unsafeInterleaveIO $ do
lock <- tryLock fp
case lock of
Nothing -> return Nothing
Just lock_fd -> return . Just $ (fp, lock_fd)
-- Attempt to rotate all files from src folder to dst folder.
rotate :: FilePath -> FilePath -> IO ()
rotate src dst = do
candidates <- getAbsoluteDirectoryFiles src
unless (null candidates)
(mapM_ doMove candidates)
where
doMove src_file = do
(new_path, h) <- mkstemp dst
hClose h
renameFile src_file new_path
getAbsoluteDirectoryFiles :: FilePath -> IO [FilePath]
getAbsoluteDirectoryFiles =
getAbsoluteDirectoryContents >=> filterM doesFileExist
getAbsoluteDirectoryContents :: FilePath -> IO [FilePath]
getAbsoluteDirectoryContents fp =
map (\rel -> joinPath [fp, rel]) <$> getDirectoryContents fp
doAppend :: FilePath -> ByteString -> IO ()
doAppend = S.appendFile
newPointsDir :: SpoolName -> FilePath
newPointsDir = specificSpoolDir ["points", "new"]
newContentsDir :: SpoolName -> FilePath
newContentsDir = specificSpoolDir ["contents", "new"]
curPointsDir :: SpoolName -> FilePath
curPointsDir = specificSpoolDir ["points", "cur"]
curContentsDir :: SpoolName -> FilePath
curContentsDir = specificSpoolDir ["contents", "cur"]
specificSpoolDir :: [String] -> SpoolName -> FilePath
specificSpoolDir subdirs sn =
addTrailingPathSeparator (joinPath (baseSpoolDir sn:subdirs))
baseSpoolDir :: SpoolName -> FilePath
baseSpoolDir (SpoolName sn) = joinPath ["/var/spool/marquise/", sn]
| anchor/marquise | lib/Marquise/IO/SpoolFile.hs | bsd-3-clause | 4,886 | 0 | 16 | 1,020 | 1,079 | 571 | 508 | 98 | 2 |
module Blockchain.SampleTransactions where
import Prelude hiding (EQ)
import qualified Data.ByteString as B
import Network.Haskoin.Internals hiding (Address)
import Blockchain.Data.Address
import Blockchain.Data.Code
import Blockchain.Data.Transaction
import Blockchain.Constants
import Blockchain.ExtendedECDSA
import Blockchain.ExtWord
import Blockchain.JCommand
import Blockchain.VM.Code
import Blockchain.VM.Labels
import Blockchain.VM.Opcodes
--import Debug.Trace
createContract::Monad m=>Integer->Integer->Code->PrvKey->SecretT m Transaction
createContract val gl code prvKey =
createContractCreationTX 0 0x9184e72a000 gl val code prvKey
createMessage::Monad m=>Integer->Integer->Address->B.ByteString->PrvKey->SecretT m Transaction
createMessage val gl toAddr theData prvKey = createMessageTX 0 0x9184e72a000 gl toAddr val theData prvKey
----------------------
simpleTX::Monad m=>PrvKey->SecretT m Transaction
simpleTX =
createContract 0 550
$ compile
[
PUSH [2],
PUSH [0],
MSTORE,
PUSH [0x20],
PUSH [0],
RETURN
]
outOfGasTX::Monad m=>PrvKey->SecretT m Transaction
outOfGasTX =
createContract 3 522
$ compile
[
PUSH [1],
PUSH [0],
MSTORE
]
simpleStorageTX::Monad m=>PrvKey->SecretT m Transaction
simpleStorageTX =
createContract 3 1000
$ compile
[
PUSH [1],
PUSH [0],
SSTORE
]
createInit::[JCommand]->[JCommand]->Code
createInit initFunc contract = -- trace (intercalate "-" $ show <$> contract) $
-- trace (intercalate "\n " $ fmap show $ snd $ jcompile $ initFunc ++ [ReturnCode contract]) $ do
compile $ lcompile $ snd $ jcompile $ initFunc ++ [ReturnCode $ compile $ lcompile $ snd $ jcompile contract]
createContractTX::Monad m=>PrvKey->SecretT m Transaction
createContractTX =
createContract (1000*finney) 1000
$ createInit []
[
PermStorage (Number 0) :=: Input 0
]
sendMessageTX::Monad m=>PrvKey->SecretT m Transaction
sendMessageTX =
createMessage (1000*finney) 1000 (Address 0x9f840fe058ce3d84e319b8c747accc1e52f69426)
(B.pack $ word256ToBytes 0x1234)
paymentContract::Monad m=>PrvKey->SecretT m Transaction
paymentContract =
createContract (1000*finney) 2000
$ createInit
[
PermStorage Caller :=: Number 1000
]
(
let
toAddr = Input (0*32)
fromAddr = Caller
val = Input (1*32)
in
[
If (PermVal fromAddr :>=: val)
[
PermStorage fromAddr :=: PermVal fromAddr - val,
PermStorage toAddr :=: PermVal toAddr + val
]
]
)
sendCoinTX::Monad m=>PrvKey->SecretT m Transaction
sendCoinTX =
createMessage 0 2000 (Address 0x9f840fe058ce3d84e319b8c747accc1e52f69426)
(B.pack $ word256ToBytes 0x1 ++ word256ToBytes 500)
keyValuePublisher::Monad m=>PrvKey->SecretT m Transaction
keyValuePublisher =
createContract (1000*finney) 2000
$ createInit
[
PermStorage 69 :=: Caller
]
(
let
inputP = MemStorage (Number 0)
inputPr = MemVal (Number 0)
in
[
If (Caller :==: PermVal (Number 69))
[
While (inputPr :<: CallDataSize)
[
PermStorage (Input inputPr) :=: Input (inputPr + 32),
inputP :=: inputPr + 64
]
]
]
)
sendKeyVal::Monad m=>PrvKey->SecretT m Transaction
sendKeyVal prvKey =
createMessage 0 2000 (Address 0x9f840fe058ce3d84e319b8c747accc1e52f69426)
(B.pack $ word256ToBytes 1000 ++ word256ToBytes 2000 ++ word256ToBytes 1234 ++ word256ToBytes 1)
prvKey
{-
(when (= (caller) @@69)
(for {} (< @i (calldatasize)) [i](+ @i 64)
[[ (calldataload @i) ]] (calldataload (+ @i 32))
)
-}
mysteryCode::[Operation]
mysteryCode =
[
PUSH [3,144],
DUP1,
PUSH [0,14],
PUSH [0],
CODECOPY,
PUSH [3,158],
JUMP,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [32],
MSTORE,
PUSH [0],
PUSH [64],
MSTORE,
PUSH [1],
PUSH [96],
MSTORE,
PUSH [2],
PUSH [128],
MSTORE,
PUSH [3],
PUSH [160],
MSTORE,
PUSH [0],
PUSH [192],
MSTORE,
PUSH [1],
PUSH [224],
MSTORE,
PUSH [2],
PUSH [1,0],
MSTORE,
PUSH [0],
PUSH [1,32],
MSTORE,
PUSH [1],
PUSH [1,64],
MSTORE,
PUSH [2],
PUSH [1,96],
MSTORE,
PUSH [3],
PUSH [1,128],
MSTORE,
PUSH [3],
PUSH [1,160],
MSTORE,
PUSH [32],
CALLDATASIZE,
DIV,
PUSH [1,192],
MSTORE,
PUSH [1,160],
MLOAD,
PUSH [1,128],
MLOAD,
ADD,
PUSH [1,192],
MLOAD,
SLT,
ISZERO,
PUSH [0,136],
JUMPI,
PUSH [1,64],
MLOAD,
PUSH [1,224],
MSTORE,
PUSH [32],
PUSH [1,224],
CALLCODE,
JUMPDEST,
PUSH [0],
PUSH [1,128],
MLOAD,
PUSH [1,160],
MLOAD,
PUSH [1,192],
MLOAD,
SUB,
SMOD,
EQ,
ISZERO,
ISZERO,
PUSH [0,174],
JUMPI,
PUSH [1,64],
MLOAD,
PUSH [2,0],
MSTORE,
PUSH [32],
PUSH [2,0],
CALLCODE,
JUMPDEST,
PUSH [1,128],
MLOAD,
PUSH [1,160],
MLOAD,
PUSH [1,192],
MLOAD,
SUB,
SDIV,
PUSH [2,32],
MSTORE,
PUSH [0],
PUSH [2,64],
MSTORE,
PUSH [2],
PUSH [1,160],
MLOAD,
ADD,
PUSH [32],
MUL,
CALLDATALOAD,
PUSH [2,96],
MSTORE,
PUSH [0],
PUSH [2,128],
MSTORE,
JUMPDEST,
PUSH [2,32],
MLOAD,
PUSH [2,64],
MLOAD,
SLT,
ISZERO,
PUSH [1,155],
JUMPI,
PUSH [1],
PUSH [1,160],
MLOAD,
PUSH [1,128],
MLOAD,
PUSH [2,64],
MLOAD,
MUL,
ADD,
ADD,
PUSH [32],
MUL,
CALLDATALOAD,
PUSH [2,160],
MSTORE,
PUSH [2],
PUSH [1,160],
MLOAD,
PUSH [1,128],
MLOAD,
PUSH [2,64],
MLOAD,
MUL,
ADD,
ADD,
PUSH [32],
MUL,
CALLDATALOAD,
PUSH [2,192],
MSTORE,
PUSH [2,96],
MLOAD,
PUSH [2,192],
MLOAD,
EQ,
ISZERO,
ISZERO,
PUSH [1,80],
JUMPI,
PUSH [2,192],
MLOAD,
PUSH [2,96],
MSTORE,
PUSH [0],
PUSH [2,128],
MLOAD,
EQ,
ISZERO,
ISZERO,
PUSH [1,79],
JUMPI,
PUSH [1,96],
MLOAD,
PUSH [2,224],
MSTORE,
PUSH [32],
PUSH [2,224],
CALLCODE,
JUMPDEST,
JUMPDEST,
PUSH [2,160],
MLOAD,
PUSH [2,128],
MLOAD,
ADD,
PUSH [2,128],
MSTORE,
PUSH [1],
PUSH [2,32],
MLOAD,
SUB,
PUSH [2,64],
MLOAD,
EQ,
ISZERO,
PUSH [1,139],
JUMPI,
PUSH [0],
PUSH [2,128],
MLOAD,
EQ,
ISZERO,
ISZERO,
PUSH [1,138],
JUMPI,
PUSH [1,96],
MLOAD,
PUSH [3,0],
MSTORE,
PUSH [32],
PUSH [3,0],
CALLCODE,
JUMPDEST,
JUMPDEST,
PUSH [1],
PUSH [2,64],
MLOAD,
ADD,
PUSH [2,64],
MSTORE,
PUSH [0,220],
JUMP,
JUMPDEST,
PUSH [32],
MLOAD,
SLOAD,
PUSH [3,32],
MSTORE,
PUSH [1],
PUSH [32],
MLOAD,
SLOAD,
ADD,
PUSH [32],
MLOAD,
SSTORE,
PUSH [32],
CALLDATALOAD,
PUSH [3,64],
MSTORE,
PUSH [64],
CALLDATALOAD,
PUSH [3,96],
MSTORE,
PUSH [255,255,255,255,255,255,255,255],
PUSH [3,128],
MSTORE,
PUSH [3,64],
MLOAD,
PUSH [64],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [3,128],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
PUSH [3,96],
MLOAD,
PUSH [96],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [3,128],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
NUMBER,
PUSH [128],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [3,128],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
TIMESTAMP,
PUSH [160],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [3,128],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
PUSH [0],
PUSH [2,64],
MSTORE,
JUMPDEST,
PUSH [2,32],
MLOAD,
PUSH [2,64],
MLOAD,
SLT,
ISZERO,
PUSH [3,129],
JUMPI,
PUSH [1,160],
MLOAD,
PUSH [1,128],
MLOAD,
PUSH [2,64],
MLOAD,
MUL,
ADD,
PUSH [32],
MUL,
CALLDATALOAD,
PUSH [3,160],
MSTORE,
PUSH [1],
PUSH [1,160],
MLOAD,
PUSH [1,128],
MLOAD,
PUSH [2,64],
MLOAD,
MUL,
ADD,
ADD,
PUSH [32],
MUL,
CALLDATALOAD,
PUSH [2,160],
MSTORE,
PUSH [2],
PUSH [1,160],
MLOAD,
PUSH [1,128],
MLOAD,
PUSH [2,64],
MLOAD,
MUL,
ADD,
ADD,
PUSH [32],
MUL,
CALLDATALOAD,
PUSH [2,192],
MSTORE,
PUSH [3,160],
MLOAD,
PUSH [192],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [2,64],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
PUSH [2,160],
MLOAD,
PUSH [224],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [2,64],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
PUSH [2,192],
MLOAD,
PUSH [1,0],
MLOAD,
PUSH [1,0,0,0,0,0,0,0,0],
PUSH [2,64],
MLOAD,
MUL,
PUSH [1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],
PUSH [3,32],
MLOAD,
MUL,
ADD,
ADD,
SSTORE,
PUSH [1],
PUSH [2,64],
MLOAD,
ADD,
PUSH [2,64],
MSTORE,
PUSH [2,138],
JUMP,
JUMPDEST,
PUSH [1,32],
MLOAD,
PUSH [3,192],
MSTORE,
PUSH [32],
PUSH [3,192],
CALLCODE,
JUMPDEST,
PUSH [0],
CALLCODE
]
createMysteryContract::Monad m=>PrvKey->SecretT m Transaction
createMysteryContract prvKey =
createContractCreationTX 0 0x9184e72a000 8000 0 (compile mysteryCode) prvKey
| jamshidh/ethereum-vm | src/Blockchain/SampleTransactions.hs | bsd-3-clause | 11,275 | 0 | 18 | 4,373 | 4,503 | 2,637 | 1,866 | 537 | 1 |
{-# LANGUAGE GeneralizedNewtypeDeriving,
FlexibleInstances,
MultiParamTypeClasses,
TypeFamilies,
StandaloneDeriving #-}
-----------------------------------------------------------------------------
-- |
-- Module : Network.Kontiki.Monad
-- Copyright : (c) 2013, Nicolas Trangez
-- License : BSD-like
--
-- Maintainer : [email protected]
--
-- This module introduces the main monad transformer `TransitionT' in
-- which `Handler's, `MessageHandler's and `TimeoutHandler's live
-- and defines actions that can be performed by them.
-----------------------------------------------------------------------------
module Network.Kontiki.Monad where
import Prelude hiding (log)
import Control.Monad.RWS
import Data.ByteString (ByteString)
import Data.ByteString.Lazy.Builder (Builder, byteString)
import Control.Lens hiding (Index)
import Network.Kontiki.Log
import Network.Kontiki.Types
-- | kontiki monad in which `Handler's,
-- `MessageHandler's and `TimeoutHandler's live that adds the ability to
-- read `Config', issue `Command's and keep state `s' to an inner monad `m'.
newtype TransitionT a s m r = TransitionT { unTransitionT :: RWST Config [Command a] s m r }
deriving ( Functor
, Applicative
, Monad
, MonadReader Config
, MonadWriter [Command a]
, MonadState s
, MonadRWS Config [Command a] s
, MonadTrans
)
instance (Monad m, MonadLog m a) => MonadLog (TransitionT a f m) a where
logEntry = lift . logEntry
logLastEntry = lift logLastEntry
-- | Runs `TransitionT'.
runTransitionT :: TransitionT a s m r -> Config -> s -> m (r, s, [Command a])
runTransitionT = runRWST . unTransitionT
-- | Broadcasts a message `m' to all nodes.
broadcast :: (Monad m, IsMessage t a) => t -> TransitionT a f m ()
broadcast m = tell [CBroadcast $ toMessage m]
-- | Sends a message `m' to a specific `NodeId' `n'.
send :: (Monad m, IsMessage t a) => NodeId -> t -> TransitionT a f m ()
send n m = tell [CSend n (toMessage m)]
-- | Resets the election timeout.
resetElectionTimeout :: Monad m => TransitionT a f m ()
resetElectionTimeout = do
t <- view configElectionTimeout
tell [CResetElectionTimeout t (2 * t)]
-- | Resets the heartbeat timeout.
resetHeartbeatTimeout :: Monad m => TransitionT a f m ()
resetHeartbeatTimeout = do
t <- view configHeartbeatTimeout
tell [CResetHeartbeatTimeout t]
-- | Logs a message from this `Builder' `b'.
log :: Monad m => Builder -> TransitionT a f m ()
log b = tell [CLog b]
-- | Logs a message from this `ByteString'.
logS :: Monad m => ByteString -> TransitionT a f m ()
logS = log . byteString
-- | Truncates the log of events to `Index' `i'.
truncateLog :: Monad m => Index -> TransitionT a f m ()
truncateLog i = tell [CTruncateLog i]
-- | Adds entries `es' to the log.
logEntries :: Monad m => [Entry a] -> TransitionT a f m ()
logEntries es = tell [CLogEntries es]
-- | Sets new commit `Index' `i'
setCommitIndex :: Monad m => Index -> TransitionT a f m ()
setCommitIndex i = tell [CSetCommitIndex i]
-- | Handler of events.
type Handler a s m =
Event a -- ^ `Event' to handle
-> TransitionT a (InternalState s) m SomeState -- ^ new `TransitionT'
-- | Handler of messages.
type MessageHandler t a s m =
NodeId -- ^ sender of the message
-> t -- ^ the mesage
-> TransitionT a (InternalState s) m SomeState -- ^ new `TransitionT'
-- | Handler of timeouts.
type TimeoutHandler t a s m = TransitionT a (InternalState s) m SomeState
| NicolasT/kontiki | src/Network/Kontiki/Monad.hs | bsd-3-clause | 3,757 | 0 | 11 | 945 | 846 | 457 | 389 | 57 | 1 |
{-
A simple homogeneous pair type with useful Functor, Applicative, and
Traversable instances.
-}
{-# LANGUAGE CPP #-}
module Pair ( Pair(..), unPair, toPair, swap ) where
#include "HsVersions.h"
import Outputable
import Control.Applicative
#if __GLASGOW_HASKELL__ < 709
import Data.Foldable
import Data.Monoid
import Data.Traversable
#endif
data Pair a = Pair { pFst :: a, pSnd :: a }
-- Note that Pair is a *unary* type constructor
-- whereas (,) is binary
-- The important thing about Pair is that it has a *homogenous*
-- Functor instance, so you can easily apply the same function
-- to both components
instance Functor Pair where
fmap f (Pair x y) = Pair (f x) (f y)
instance Applicative Pair where
pure x = Pair x x
(Pair f g) <*> (Pair x y) = Pair (f x) (g y)
instance Foldable Pair where
foldMap f (Pair x y) = f x `mappend` f y
instance Traversable Pair where
traverse f (Pair x y) = Pair <$> f x <*> f y
instance Outputable a => Outputable (Pair a) where
ppr (Pair a b) = ppr a <+> char '~' <+> ppr b
unPair :: Pair a -> (a,a)
unPair (Pair x y) = (x,y)
toPair :: (a,a) -> Pair a
toPair (x,y) = Pair x y
swap :: Pair a -> Pair a
swap (Pair x y) = Pair y x
| green-haskell/ghc | compiler/utils/Pair.hs | bsd-3-clause | 1,191 | 0 | 8 | 253 | 430 | 227 | 203 | 25 | 1 |
{-# language DataKinds #-}
{-# language OverloadedLists #-}
{-# language OverloadedStrings #-}
{-# language QuasiQuotes #-}
{-# language Rank2Types #-}
{-# language TypeApplications #-}
{-# language TypeFamilies #-}
module Planetary.Core.Eval.Test (unitTests, runTest, mkLogger) where
import Control.Lens
import Control.Monad.Reader (asks)
import Control.Monad.IO.Class
import Data.Maybe (fromJust)
import Data.Text (Text)
import NeatInterpolation
import Network.IPLD (Cid, toIpld)
import Prelude hiding (not)
import EasyTest hiding (bool, run)
import Planetary.Core hiding (logIncomplete, logReturnState, logValue)
import Planetary.Support.Ids hiding (boolId) -- XXX fix this
import Planetary.Support.Parser (forceTm)
import Planetary.Support.Pretty
import Planetary.Library
import qualified Planetary.Library.FrankExamples as Frank
import Planetary.Library.HaskellForeign (mkForeignTm, haskellOracles, intOpsId)
import Data.Text.Prettyprint.Doc
-- Some more good examples here:
-- https://github.com/dhil/links/blob/master/examples/handlers/shallow_state.links
noteFailureState
:: EvalState -> Either Err TmI -> Either Err TmI -> Test ()
noteFailureState initState result expected = do
note $ layout $ vsep
[ ""
, annotate Error "fail with initial state:"
, prettyEvalState initState
, ""
, annotate Error "got:"
, either pretty (prettyTmPrec 11) result
, ""
, annotate Error "expected:"
, either pretty (prettyTmPrec 11) expected
]
fail "failure: see above"
putLogs :: Bool
putLogs = False
mkLogger :: (Text -> IO ()) -> Logger
mkLogger mkNote =
let helper :: forall a. (a -> Text) -> a -> IO ()
helper f = if putLogs then mkNote . f else const (pure ())
in Logger (helper . logReturnState) (helper logIncomplete) (helper logValue)
runTest
:: Text
-> AmbientHandlers
-> ValueStore
-> TmI
-> Either Err TmI
-> Test ()
runTest name handlers store tm expected = scope name $ do
let initState = initEvalState store tm
logger <- mkLogger <$> asks note_
result <- liftIO $ run handlers logger initState
if result == expected
then ok
else noteFailureState initState result expected
boolId :: Cid
(boolId, _) = fromJust $ namedData "Bool" Frank.resolvedDecls
bool :: Int -> TmI
bool i = DataConstructor boolId i []
unitTests :: Test ()
unitTests =
let
-- true, false :: forall a b. Tm Cid a b
false = bool 0
true = bool 1
not tm = Case tm
[ ([], true)
, ([], false)
]
evalEnvRunTest desc = runTest desc noAmbientHandlers emptyStore
in scope "evaluation" $ tests
[ let x = V"x"
-- tm = forceTm "(\y -> y) x"
lam = Lambda ["x"] x
in scope "functions" $ tests
-- [ evalEnvRunTest "application 1" (AppN lam [true])
-- (Right true)
[ evalEnvRunTest "application"
(AppT lam [true])
(Right true)
-- TODO: test further steps with bound variables
, evalEnvRunTest "nullary function call"
(AppT (Lambda [] true) [])
(Right true)
]
, scope "case" $ tests
[ evalEnvRunTest "not (1)" (not false) (Right true)
, evalEnvRunTest "not (2)" (not true) (Right false)
]
, let ty :: Polytype Cid
ty = Polytype [] (DataTy (UidTy boolId) [])
tm = Let false ty "x" (V"x")
in scope "let" $ evalEnvRunTest "let x = false in x" tm (Right false)
, scope "handle" $ do
let abortHandlerTm = forceTm [text|
handle x : [e , <Abort>]Int with
Abort:
| <aborting -> k> -> one
| v -> two
|]
sendHandlerTm = forceTm [text|
handle x : [e, <Send Int>]Int with
Send:
| <send n -> k> -> n
| v -> v
|]
-- TODO: this is duplicated in FrankExamples.Test
stateHandlerTm = forceTm [text|
letrec
state : forall S X. {S -> {X} -> X}
= \s x -> handle x! : X with
State:
| <get -> k> -> state s (\-> k s)
| <put s -> k> -> state s (\-> k <Unit.0>)
| y -> y
fst : forall X Y. {X -> Y -> X}
= \x y -> x
-- increment, return original value
next : forall. {[<State Int>]Int}
-- fst get! (put (get! + 1))
= \-> fst get! (put (add get! one))
statefulTm
: forall. {[<State Int>] Int}
= \-> let x : forall. Int = next! in
let y : forall. Int = next! in
let z : forall. Int = next! in z
in state zero statefulTm
|]
(zero, zeroVal) = mkForeignTm @Int intId [] 0
(one, oneVal) = mkForeignTm @Int intId [] 1
(two, twoVal) = mkForeignTm @Int intId [] 2
Right abortHandler <- pure $ resolve $ fst abortHandlerTm
Right sendHandler <- pure $ resolve $ fst sendHandlerTm
Right stateHandler <- pure $ resolve $ fst stateHandlerTm
Just abortCid <- pure $ Frank.resolvedDecls ^? globalCids . ix "Abort"
Just sendCid <- pure $ Frank.resolvedDecls ^? globalCids . ix "Send"
Just stateCid <- pure $ Frank.resolvedDecls ^? globalCids . ix "State"
let abortHandler' = substitute "one" one $
substitute "two" two
abortHandler
handleVal = substitute "x" zero abortHandler'
abort = AppN (Command abortCid 0) []
handleAbort = substitute "x" abort abortHandler'
handleSend = substitute "x"
(AppN (Command sendCid 0) [one])
sendHandler
get = Command stateCid 0
put = Command stateCid 1
add = Command intOpsId 0
handleNext = substituteAll
[ ("zero", zero)
, ("one", one)
, ("get", get)
, ("put", put)
, ("add", add)
]
stateHandler
numberStore = storeOf $ toIpld <$> [zeroVal, oneVal, twoVal]
tests
[ runTest "val" noAmbientHandlers emptyStore handleVal
(Right two)
, runTest "abort" noAmbientHandlers emptyStore
handleAbort (Right one)
, runTest "send" noAmbientHandlers emptyStore handleSend
(Right one)
, let handlers = AmbientHandlers haskellOracles
in runTest "handle state" handlers numberStore handleNext (Right two)
]
, scope "let x = false in let y = not x in not y" $ do
let
ty = Polytype [] (DataTy (UidTy boolId) [])
tm = Let false ty "x" $
Let (not (V"x")) ty "y" $
not (V"y")
-- both versions of tm should be equivalent
Right tm2 <- pure $ resolve $ fst $ forceTm [text|
let not: forall. {<Bool> -> <Bool>}
= \x -> case x of
| <False> -> <Bool.1>
| <True> -> <Bool.0>
in
let x: forall. Bool = false in
let y: forall. Bool = not x in
not y
|]
let tm2' = substitute "false" false tm2
tests
[ evalEnvRunTest "tm" tm (Right false)
, evalEnvRunTest "tm2" tm2' (Right false)
]
, scope "letrec" $ do
let evenodd = fst $ forceTm [text|
letrec
even : forall. {<Fix NatF> -> <Bool>}
= \n -> case n of
| <z> -> <Bool.1> -- true
| <succ n'> -> odd n'
odd : forall. {<Fix NatF> -> <Bool>}
= \m -> case m of
| <z> -> <Bool.0> -- false
| <succ m'> -> even m'
in body
|]
-- mkFix = Command fixOpsId 0
-- unFix = Command fixOpsId 1
Right evenodd' <- pure $ resolve evenodd
Just (natfId, _) <- pure $ namedData "NatF" Frank.resolvedDecls
let -- mkTm n = [| evenOdd n |]
mkTm :: Text -> Int -> TmI
mkTm fnName n =
let mkNat 0 = DataConstructor natfId 0 []
mkNat k = DataConstructor natfId 1 [mkNat (k - 1)]
tm = substitute "body"
(AppT (V fnName) [mkNat n])
evenodd'
in tm
handlers = AmbientHandlers haskellOracles
runTest' desc = runTest desc handlers emptyStore
tests
[ runTest' "even 0" (mkTm "even" 0) (Right true)
, runTest' "odd 0" (mkTm "odd" 0) (Right false)
, runTest' "even 1" (mkTm "even" 1) (Right false)
, runTest' "odd 1" (mkTm "odd" 1) (Right true)
, runTest' "even 7" (mkTm "even" 7) (Right false)
, runTest' "odd 7" (mkTm "odd" 7) (Right true)
, runTest' "even 10" (mkTm "even" 10) (Right true)
, runTest' "odd 10" (mkTm "odd" 10) (Right false)
, runTest' "odd 20" (mkTm "odd" 20) (Right false)
, runTest' "even 100" (mkTm "even" 100) (Right true)
]
, scope "closures" $ do
let tm = fst $ forceTm [text|
letrec
const
: forall. {<Text> -> {<Text> -> <Text>}}
= \x -> \y -> x
-- capture x, then shadow
actual1
: forall. {<Text>}
= \->
let foo' : forall. {<Text> -> <Text>} = const foo in
let x : forall. <Text> = bar in
foo' bar
expected1
: forall. {<Text>}
= \-> foo
in actual1!
|]
let tm2 = fst $ forceTm [text|
letrec
captureX
: forall. {<Text> -> <Pair {<Text> -> <Text>} <Text>>}
= \x -> <Pair.0 (\y -> x) x>
-- capture x, then shadow
actual1
: forall. {<Text>}
= \->
let pair : forall. <Pair {<Text> -> <Text>} <Text>> = captureX foo in
case pair of
| <pair f x> -> f bar
expected1
: forall. {<Text>}
= \-> foo
in actual1!
|]
Right tm' <- pure $ resolve tm
Right tm2' <- pure $ resolve tm2
let (foo, fooVal) = mkForeignTm @Text intId [] "foo"
(bar, barVal) = mkForeignTm @Text intId [] "bar"
(baz, bazVal) = mkForeignTm @Text intId [] "baz"
subs =
[ ("foo", foo)
, ("bar", bar)
, ("baz", baz)
]
let tm'' = substituteAll subs tm'
tm2'' = substituteAll subs tm2'
store = storeOf $ toIpld <$> [fooVal, barVal, bazVal]
tests
[ runTest "const" noAmbientHandlers store tm'' (Right foo)
, runTest "pair" noAmbientHandlers store tm2'' (Right foo)
]
]
| joelburget/interplanetary-computation | src/Planetary/Core/Eval/Test.hs | bsd-3-clause | 11,459 | 0 | 24 | 4,553 | 2,481 | 1,275 | 1,206 | -1 | -1 |
{-# LANGUAGE CPP, MagicHash #-}
-- | Dynamically lookup up values from modules and loading them.
module DynamicLoading (
#ifdef GHCI
-- * Loading plugins
loadPlugins,
loadFrontendPlugin,
-- * Force loading information
forceLoadModuleInterfaces,
forceLoadNameModuleInterface,
forceLoadTyCon,
-- * Finding names
lookupRdrNameInModuleForPlugins,
-- * Loading values
getValueSafely,
getHValueSafely,
lessUnsafeCoerce
#else
pluginError,
#endif
) where
#ifdef GHCI
import Linker ( linkModule, getHValue )
import GHCi ( wormhole )
import SrcLoc ( noSrcSpan )
import Finder ( findPluginModule, cannotFindModule )
import TcRnMonad ( initTcInteractive, initIfaceTcRn )
import LoadIface ( loadPluginInterface )
import RdrName ( RdrName, ImportSpec(..), ImpDeclSpec(..)
, ImpItemSpec(..), mkGlobalRdrEnv, lookupGRE_RdrName
, gre_name, mkRdrQual )
import OccName ( OccName, mkVarOcc )
import RnNames ( gresFromAvails )
import DynFlags
import Plugins ( Plugin, FrontendPlugin, CommandLineOption )
import PrelNames ( pluginTyConName, frontendPluginTyConName )
import HscTypes
import GHCi.RemoteTypes ( HValue )
import Type ( Type, eqType, mkTyConTy, pprTyThingCategory )
import TyCon ( TyCon )
import Name ( Name, nameModule_maybe )
import Id ( idType )
import Module ( Module, ModuleName )
import Panic
import FastString
import ErrUtils
import Outputable
import Exception
import Hooks
import Data.Maybe ( mapMaybe )
import GHC.Exts ( unsafeCoerce# )
#else
import Module ( ModuleName, moduleNameString )
import Panic
import Data.List ( intercalate )
#endif
#ifdef GHCI
loadPlugins :: HscEnv -> IO [(ModuleName, Plugin, [CommandLineOption])]
loadPlugins hsc_env
= do { plugins <- mapM (loadPlugin hsc_env) to_load
; return $ map attachOptions $ to_load `zip` plugins }
where
dflags = hsc_dflags hsc_env
to_load = pluginModNames dflags
attachOptions (mod_nm, plug) = (mod_nm, plug, options)
where
options = [ option | (opt_mod_nm, option) <- pluginModNameOpts dflags
, opt_mod_nm == mod_nm ]
loadPlugin :: HscEnv -> ModuleName -> IO Plugin
loadPlugin = loadPlugin' (mkVarOcc "plugin") pluginTyConName
loadFrontendPlugin :: HscEnv -> ModuleName -> IO FrontendPlugin
loadFrontendPlugin = loadPlugin' (mkVarOcc "frontendPlugin") frontendPluginTyConName
loadPlugin' :: OccName -> Name -> HscEnv -> ModuleName -> IO a
loadPlugin' occ_name plugin_name hsc_env mod_name
= do { let plugin_rdr_name = mkRdrQual mod_name occ_name
dflags = hsc_dflags hsc_env
; mb_name <- lookupRdrNameInModuleForPlugins hsc_env mod_name
plugin_rdr_name
; case mb_name of {
Nothing ->
throwGhcExceptionIO (CmdLineError $ showSDoc dflags $ hsep
[ text "The module", ppr mod_name
, text "did not export the plugin name"
, ppr plugin_rdr_name ]) ;
Just name ->
do { plugin_tycon <- forceLoadTyCon hsc_env plugin_name
; mb_plugin <- getValueSafely hsc_env name (mkTyConTy plugin_tycon)
; case mb_plugin of
Nothing ->
throwGhcExceptionIO (CmdLineError $ showSDoc dflags $ hsep
[ text "The value", ppr name
, text "did not have the type"
, ppr pluginTyConName, text "as required"])
Just plugin -> return plugin } } }
-- | Force the interfaces for the given modules to be loaded. The 'SDoc' parameter is used
-- for debugging (@-ddump-if-trace@) only: it is shown as the reason why the module is being loaded.
forceLoadModuleInterfaces :: HscEnv -> SDoc -> [Module] -> IO ()
forceLoadModuleInterfaces hsc_env doc modules
= (initTcInteractive hsc_env $
initIfaceTcRn $
mapM_ (loadPluginInterface doc) modules)
>> return ()
-- | Force the interface for the module containing the name to be loaded. The 'SDoc' parameter is used
-- for debugging (@-ddump-if-trace@) only: it is shown as the reason why the module is being loaded.
forceLoadNameModuleInterface :: HscEnv -> SDoc -> Name -> IO ()
forceLoadNameModuleInterface hsc_env reason name = do
let name_modules = mapMaybe nameModule_maybe [name]
forceLoadModuleInterfaces hsc_env reason name_modules
-- | Load the 'TyCon' associated with the given name, come hell or high water. Fails if:
--
-- * The interface could not be loaded
-- * The name is not that of a 'TyCon'
-- * The name did not exist in the loaded module
forceLoadTyCon :: HscEnv -> Name -> IO TyCon
forceLoadTyCon hsc_env con_name = do
forceLoadNameModuleInterface hsc_env (text "contains a name used in an invocation of loadTyConTy") con_name
mb_con_thing <- lookupTypeHscEnv hsc_env con_name
case mb_con_thing of
Nothing -> throwCmdLineErrorS dflags $ missingTyThingError con_name
Just (ATyCon tycon) -> return tycon
Just con_thing -> throwCmdLineErrorS dflags $ wrongTyThingError con_name con_thing
where dflags = hsc_dflags hsc_env
-- | Loads the value corresponding to a 'Name' if that value has the given 'Type'. This only provides limited safety
-- in that it is up to the user to ensure that that type corresponds to the type you try to use the return value at!
--
-- If the value found was not of the correct type, returns @Nothing@. Any other condition results in an exception:
--
-- * If we could not load the names module
-- * If the thing being loaded is not a value
-- * If the Name does not exist in the module
-- * If the link failed
getValueSafely :: HscEnv -> Name -> Type -> IO (Maybe a)
getValueSafely hsc_env val_name expected_type = do
mb_hval <- lookupHook getValueSafelyHook getHValueSafely dflags hsc_env val_name expected_type
case mb_hval of
Nothing -> return Nothing
Just hval -> do
value <- lessUnsafeCoerce dflags "getValueSafely" hval
return (Just value)
where
dflags = hsc_dflags hsc_env
getHValueSafely :: HscEnv -> Name -> Type -> IO (Maybe HValue)
getHValueSafely hsc_env val_name expected_type = do
forceLoadNameModuleInterface hsc_env (text "contains a name used in an invocation of getHValueSafely") val_name
-- Now look up the names for the value and type constructor in the type environment
mb_val_thing <- lookupTypeHscEnv hsc_env val_name
case mb_val_thing of
Nothing -> throwCmdLineErrorS dflags $ missingTyThingError val_name
Just (AnId id) -> do
-- Check the value type in the interface against the type recovered from the type constructor
-- before finally casting the value to the type we assume corresponds to that constructor
if expected_type `eqType` idType id
then do
-- Link in the module that contains the value, if it has such a module
case nameModule_maybe val_name of
Just mod -> do linkModule hsc_env mod
return ()
Nothing -> return ()
-- Find the value that we just linked in and cast it given that we have proved it's type
hval <- getHValue hsc_env val_name >>= wormhole dflags
return (Just hval)
else return Nothing
Just val_thing -> throwCmdLineErrorS dflags $ wrongTyThingError val_name val_thing
where dflags = hsc_dflags hsc_env
-- | Coerce a value as usual, but:
--
-- 1) Evaluate it immediately to get a segfault early if the coercion was wrong
--
-- 2) Wrap it in some debug messages at verbosity 3 or higher so we can see what happened
-- if it /does/ segfault
lessUnsafeCoerce :: DynFlags -> String -> a -> IO b
lessUnsafeCoerce dflags context what = do
debugTraceMsg dflags 3 $ (text "Coercing a value in") <+> (text context) <>
(text "...")
output <- evaluate (unsafeCoerce# what)
debugTraceMsg dflags 3 (text "Successfully evaluated coercion")
return output
-- | Finds the 'Name' corresponding to the given 'RdrName' in the
-- context of the 'ModuleName'. Returns @Nothing@ if no such 'Name'
-- could be found. Any other condition results in an exception:
--
-- * If the module could not be found
-- * If we could not determine the imports of the module
--
-- Can only be used for looking up names while loading plugins (and is
-- *not* suitable for use within plugins). The interface file is
-- loaded very partially: just enough that it can be used, without its
-- rules and instances affecting (and being linked from!) the module
-- being compiled. This was introduced by 57d6798.
lookupRdrNameInModuleForPlugins :: HscEnv -> ModuleName -> RdrName -> IO (Maybe Name)
lookupRdrNameInModuleForPlugins hsc_env mod_name rdr_name = do
-- First find the package the module resides in by searching exposed packages and home modules
found_module <- findPluginModule hsc_env mod_name
case found_module of
Found _ mod -> do
-- Find the exports of the module
(_, mb_iface) <- initTcInteractive hsc_env $
initIfaceTcRn $
loadPluginInterface doc mod
case mb_iface of
Just iface -> do
-- Try and find the required name in the exports
let decl_spec = ImpDeclSpec { is_mod = mod_name, is_as = mod_name
, is_qual = False, is_dloc = noSrcSpan }
imp_spec = ImpSpec decl_spec ImpAll
env = mkGlobalRdrEnv (gresFromAvails (Just imp_spec) (mi_exports iface))
case lookupGRE_RdrName rdr_name env of
[gre] -> return (Just (gre_name gre))
[] -> return Nothing
_ -> panic "lookupRdrNameInModule"
Nothing -> throwCmdLineErrorS dflags $ hsep [text "Could not determine the exports of the module", ppr mod_name]
err -> throwCmdLineErrorS dflags $ cannotFindModule dflags mod_name err
where
dflags = hsc_dflags hsc_env
doc = text "contains a name used in an invocation of lookupRdrNameInModule"
wrongTyThingError :: Name -> TyThing -> SDoc
wrongTyThingError name got_thing = hsep [text "The name", ppr name, ptext (sLit "is not that of a value but rather a"), pprTyThingCategory got_thing]
missingTyThingError :: Name -> SDoc
missingTyThingError name = hsep [text "The name", ppr name, ptext (sLit "is not in the type environment: are you sure it exists?")]
throwCmdLineErrorS :: DynFlags -> SDoc -> IO a
throwCmdLineErrorS dflags = throwCmdLineError . showSDoc dflags
throwCmdLineError :: String -> IO a
throwCmdLineError = throwGhcExceptionIO . CmdLineError
#else
pluginError :: [ModuleName] -> a
pluginError modnames = throwGhcException (CmdLineError msg)
where
msg = "not built for interactive use - can't load plugins ("
-- module names are not z-encoded
++ intercalate ", " (map moduleNameString modnames)
++ ")"
#endif
| vTurbine/ghc | compiler/main/DynamicLoading.hs | bsd-3-clause | 11,451 | 1 | 24 | 3,117 | 1,998 | 1,041 | 957 | 11 | 1 |
<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE helpset PUBLIC "-//Sun Microsystems Inc.//DTD JavaHelp HelpSet Version 2.0//EN" "http://java.sun.com/products/javahelp/helpset_2_0.dtd">
<helpset version="2.0" xml:lang="sr-CS">
<title>JSON View</title>
<maps>
<homeID>jsonview</homeID>
<mapref location="map.jhm"/>
</maps>
<view>
<name>TOC</name>
<label>Contents</label>
<type>org.zaproxy.zap.extension.help.ZapTocView</type>
<data>toc.xml</data>
</view>
<view>
<name>Index</name>
<label>Index</label>
<type>javax.help.IndexView</type>
<data>index.xml</data>
</view>
<view>
<name>Search</name>
<label>Search</label>
<type>javax.help.SearchView</type>
<data engine="com.sun.java.help.search.DefaultSearchEngine">
JavaHelpSearch
</data>
</view>
<view>
<name>Favorites</name>
<label>Favorites</label>
<type>javax.help.FavoritesView</type>
</view>
</helpset> | thc202/zap-extensions | addOns/jsonview/src/main/javahelp/help_sr_CS/helpset_sr_CS.hs | apache-2.0 | 959 | 77 | 66 | 156 | 407 | 206 | 201 | -1 | -1 |
-- | Put common type definitions here to break recursive module dependencies.
module RegAlloc.Linear.Base (
BlockAssignment,
Loc(..),
regsOfLoc,
-- for stats
SpillReason(..),
RegAllocStats(..),
-- the allocator monad
RA_State(..),
)
where
import GhcPrelude
import RegAlloc.Linear.StackMap
import RegAlloc.Liveness
import Reg
import DynFlags
import Outputable
import Unique
import UniqFM
import UniqSupply
-- | Used to store the register assignment on entry to a basic block.
-- We use this to handle join points, where multiple branch instructions
-- target a particular label. We have to insert fixup code to make
-- the register assignments from the different sources match up.
--
type BlockAssignment freeRegs
= BlockMap (freeRegs, RegMap Loc)
-- | Where a vreg is currently stored
-- A temporary can be marked as living in both a register and memory
-- (InBoth), for example if it was recently loaded from a spill location.
-- This makes it cheap to spill (no save instruction required), but we
-- have to be careful to turn this into InReg if the value in the
-- register is changed.
-- This is also useful when a temporary is about to be clobbered. We
-- save it in a spill location, but mark it as InBoth because the current
-- instruction might still want to read it.
--
data Loc
-- | vreg is in a register
= InReg !RealReg
-- | vreg is held in a stack slot
| InMem {-# UNPACK #-} !StackSlot
-- | vreg is held in both a register and a stack slot
| InBoth !RealReg
{-# UNPACK #-} !StackSlot
deriving (Eq, Show, Ord)
instance Outputable Loc where
ppr l = text (show l)
-- | Get the reg numbers stored in this Loc.
regsOfLoc :: Loc -> [RealReg]
regsOfLoc (InReg r) = [r]
regsOfLoc (InBoth r _) = [r]
regsOfLoc (InMem _) = []
-- | Reasons why instructions might be inserted by the spiller.
-- Used when generating stats for -ddrop-asm-stats.
--
data SpillReason
-- | vreg was spilled to a slot so we could use its
-- current hreg for another vreg
= SpillAlloc !Unique
-- | vreg was moved because its hreg was clobbered
| SpillClobber !Unique
-- | vreg was loaded from a spill slot
| SpillLoad !Unique
-- | reg-reg move inserted during join to targets
| SpillJoinRR !Unique
-- | reg-mem move inserted during join to targets
| SpillJoinRM !Unique
-- | Used to carry interesting stats out of the register allocator.
data RegAllocStats
= RegAllocStats
{ ra_spillInstrs :: UniqFM [Int] }
-- | The register allocator state
data RA_State freeRegs
= RA_State
{
-- | the current mapping from basic blocks to
-- the register assignments at the beginning of that block.
ra_blockassig :: BlockAssignment freeRegs
-- | free machine registers
, ra_freeregs :: !freeRegs
-- | assignment of temps to locations
, ra_assig :: RegMap Loc
-- | current stack delta
, ra_delta :: Int
-- | free stack slots for spilling
, ra_stack :: StackMap
-- | unique supply for generating names for join point fixup blocks.
, ra_us :: UniqSupply
-- | Record why things were spilled, for -ddrop-asm-stats.
-- Just keep a list here instead of a map of regs -> reasons.
-- We don't want to slow down the allocator if we're not going to emit the stats.
, ra_spills :: [SpillReason]
, ra_DynFlags :: DynFlags }
| ezyang/ghc | compiler/nativeGen/RegAlloc/Linear/Base.hs | bsd-3-clause | 3,779 | 0 | 10 | 1,161 | 392 | 245 | 147 | 65 | 1 |
{-# LANGUAGE Trustworthy #-}
{-# LANGUAGE NoImplicitPrelude, MagicHash #-}
-----------------------------------------------------------------------------
-- |
-- Module : Numeric
-- Copyright : (c) The University of Glasgow 2002
-- License : BSD-style (see the file libraries/base/LICENSE)
--
-- Maintainer : [email protected]
-- Stability : provisional
-- Portability : portable
--
-- Odds and ends, mostly functions for reading and showing
-- 'RealFloat'-like kind of values.
--
-----------------------------------------------------------------------------
module Numeric (
-- * Showing
showSigned,
showIntAtBase,
showInt,
showHex,
showOct,
showEFloat,
showFFloat,
showGFloat,
showFFloatAlt,
showGFloatAlt,
showFloat,
floatToDigits,
-- * Reading
-- | /NB:/ 'readInt' is the \'dual\' of 'showIntAtBase',
-- and 'readDec' is the \`dual\' of 'showInt'.
-- The inconsistent naming is a historical accident.
readSigned,
readInt,
readDec,
readOct,
readHex,
readFloat,
lexDigits,
-- * Miscellaneous
fromRat,
) where
import GHC.Base
import GHC.Read
import GHC.Real
import GHC.Float
import GHC.Num
import GHC.Show
import Data.Maybe
import Text.ParserCombinators.ReadP( ReadP, readP_to_S, pfail )
import qualified Text.Read.Lex as L
-- -----------------------------------------------------------------------------
-- Reading
-- | Reads an /unsigned/ 'Integral' value in an arbitrary base.
readInt :: Num a
=> a -- ^ the base
-> (Char -> Bool) -- ^ a predicate distinguishing valid digits in this base
-> (Char -> Int) -- ^ a function converting a valid digit character to an 'Int'
-> ReadS a
readInt base isDigit valDigit = readP_to_S (L.readIntP base isDigit valDigit)
-- | Read an unsigned number in octal notation.
readOct :: (Eq a, Num a) => ReadS a
readOct = readP_to_S L.readOctP
-- | Read an unsigned number in decimal notation.
readDec :: (Eq a, Num a) => ReadS a
readDec = readP_to_S L.readDecP
-- | Read an unsigned number in hexadecimal notation.
-- Both upper or lower case letters are allowed.
readHex :: (Eq a, Num a) => ReadS a
readHex = readP_to_S L.readHexP
-- | Reads an /unsigned/ 'RealFrac' value,
-- expressed in decimal scientific notation.
readFloat :: RealFrac a => ReadS a
readFloat = readP_to_S readFloatP
readFloatP :: RealFrac a => ReadP a
readFloatP =
do tok <- L.lex
case tok of
L.Number n -> return $ fromRational $ L.numberToRational n
_ -> pfail
-- It's turgid to have readSigned work using list comprehensions,
-- but it's specified as a ReadS to ReadS transformer
-- With a bit of luck no one will use it.
-- | Reads a /signed/ 'Real' value, given a reader for an unsigned value.
readSigned :: (Real a) => ReadS a -> ReadS a
readSigned readPos = readParen False read'
where read' r = read'' r ++
(do
("-",s) <- lex r
(x,t) <- read'' s
return (-x,t))
read'' r = do
(str,s) <- lex r
(n,"") <- readPos str
return (n,s)
-- -----------------------------------------------------------------------------
-- Showing
-- | Show /non-negative/ 'Integral' numbers in base 10.
showInt :: Integral a => a -> ShowS
showInt n0 cs0
| n0 < 0 = error "Numeric.showInt: can't show negative numbers"
| otherwise = go n0 cs0
where
go n cs
| n < 10 = case unsafeChr (ord '0' + fromIntegral n) of
c@(C# _) -> c:cs
| otherwise = case unsafeChr (ord '0' + fromIntegral r) of
c@(C# _) -> go q (c:cs)
where
(q,r) = n `quotRem` 10
-- Controlling the format and precision of floats. The code that
-- implements the formatting itself is in @PrelNum@ to avoid
-- mutual module deps.
{-# SPECIALIZE showEFloat ::
Maybe Int -> Float -> ShowS,
Maybe Int -> Double -> ShowS #-}
{-# SPECIALIZE showFFloat ::
Maybe Int -> Float -> ShowS,
Maybe Int -> Double -> ShowS #-}
{-# SPECIALIZE showGFloat ::
Maybe Int -> Float -> ShowS,
Maybe Int -> Double -> ShowS #-}
-- | Show a signed 'RealFloat' value
-- using scientific (exponential) notation (e.g. @2.45e2@, @1.5e-3@).
--
-- In the call @'showEFloat' digs val@, if @digs@ is 'Nothing',
-- the value is shown to full precision; if @digs@ is @'Just' d@,
-- then at most @d@ digits after the decimal point are shown.
showEFloat :: (RealFloat a) => Maybe Int -> a -> ShowS
-- | Show a signed 'RealFloat' value
-- using standard decimal notation (e.g. @245000@, @0.0015@).
--
-- In the call @'showFFloat' digs val@, if @digs@ is 'Nothing',
-- the value is shown to full precision; if @digs@ is @'Just' d@,
-- then at most @d@ digits after the decimal point are shown.
showFFloat :: (RealFloat a) => Maybe Int -> a -> ShowS
-- | Show a signed 'RealFloat' value
-- using standard decimal notation for arguments whose absolute value lies
-- between @0.1@ and @9,999,999@, and scientific notation otherwise.
--
-- In the call @'showGFloat' digs val@, if @digs@ is 'Nothing',
-- the value is shown to full precision; if @digs@ is @'Just' d@,
-- then at most @d@ digits after the decimal point are shown.
showGFloat :: (RealFloat a) => Maybe Int -> a -> ShowS
showEFloat d x = showString (formatRealFloat FFExponent d x)
showFFloat d x = showString (formatRealFloat FFFixed d x)
showGFloat d x = showString (formatRealFloat FFGeneric d x)
-- | Show a signed 'RealFloat' value
-- using standard decimal notation (e.g. @245000@, @0.0015@).
--
-- This behaves as 'showFFloat', except that a decimal point
-- is always guaranteed, even if not needed.
--
-- /Since: 4.7.0.0/
showFFloatAlt :: (RealFloat a) => Maybe Int -> a -> ShowS
-- | Show a signed 'RealFloat' value
-- using standard decimal notation for arguments whose absolute value lies
-- between @0.1@ and @9,999,999@, and scientific notation otherwise.
--
-- This behaves as 'showFFloat', except that a decimal point
-- is always guaranteed, even if not needed.
--
-- /Since: 4.7.0.0/
showGFloatAlt :: (RealFloat a) => Maybe Int -> a -> ShowS
showFFloatAlt d x = showString (formatRealFloatAlt FFFixed d True x)
showGFloatAlt d x = showString (formatRealFloatAlt FFGeneric d True x)
-- ---------------------------------------------------------------------------
-- Integer printing functions
-- | Shows a /non-negative/ 'Integral' number using the base specified by the
-- first argument, and the character representation specified by the second.
showIntAtBase :: (Integral a, Show a) => a -> (Int -> Char) -> a -> ShowS
showIntAtBase base toChr n0 r0
| base <= 1 = error ("Numeric.showIntAtBase: applied to unsupported base " ++ show base)
| n0 < 0 = error ("Numeric.showIntAtBase: applied to negative number " ++ show n0)
| otherwise = showIt (quotRem n0 base) r0
where
showIt (n,d) r = seq c $ -- stricter than necessary
case n of
0 -> r'
_ -> showIt (quotRem n base) r'
where
c = toChr (fromIntegral d)
r' = c : r
-- | Show /non-negative/ 'Integral' numbers in base 16.
showHex :: (Integral a,Show a) => a -> ShowS
showHex = showIntAtBase 16 intToDigit
-- | Show /non-negative/ 'Integral' numbers in base 8.
showOct :: (Integral a, Show a) => a -> ShowS
showOct = showIntAtBase 8 intToDigit
| jwiegley/ghc-release | libraries/base/Numeric.hs | gpl-3.0 | 7,732 | 0 | 13 | 1,977 | 1,332 | 726 | 606 | 107 | 2 |
module ShouldFail where
-- strictness annotations on the argument to a newtype constructor
-- are not allowed.
newtype N a = T ![a]
| urbanslug/ghc | testsuite/tests/parser/should_fail/readFail008.hs | bsd-3-clause | 133 | 2 | 6 | 24 | 20 | 14 | 6 | 4 | 0 |
{-# LANGUAGE DeriveGeneric #-}
{-# LANGUAGE LambdaCase #-}
{-# LANGUAGE OverloadedStrings #-}
{-# OPTIONS_GHC -fno-warn-orphans #-}
module Network.IPFS.Types (
IPFS (..),
Endpoint (..),
Multihash (..),
Key,
Data,
Template (..),
FileHash (..),
Link (..),
Node (..),
Object (..),
ID (..),
parseMultihash
) where
import Control.Monad (ap, liftM)
import Control.Monad.Fix (MonadFix, mfix)
import Control.Monad.IO.Class (MonadIO, liftIO)
import Control.Monad.Trans.Reader (ReaderT)
import Data.Aeson (FromJSON (..), (.:), (.:?))
import qualified Data.Aeson as JSON
import Data.ByteString.Lazy (ByteString)
import Data.ByteString.Lazy.UTF8 (fromString, toString)
import qualified Data.Multihash.Base as MB
import Data.Text (unpack)
import GHC.Generics (Generic)
import qualified Network.HTTP.Conduit as HTTP
import Network.Multiaddr (Multiaddr)
import Text.ParserCombinators.ReadP (ReadP, readP_to_S, munch1)
-- | An 'Endpoint' is an IPFS node that will execute an API request
data Endpoint = Endpoint HTTP.Manager String
newtype IPFS a = IPFS { unIPFS :: ReaderT Endpoint IO a }
instance Monad IPFS where
return = IPFS . return
m >>= f = IPFS (unIPFS m >>= unIPFS . f)
instance MonadFix IPFS where
mfix f = IPFS (mfix (unIPFS . f))
instance MonadIO IPFS where
liftIO = IPFS . liftIO
instance Functor IPFS where
fmap = liftM
instance Applicative IPFS where
pure = return
(<*>) = ap
newtype Multihash = Multihash { multihash :: ByteString } deriving (Generic, Eq)
instance Show Multihash where
show = toString . MB.encode MB.Base58 . multihash
instance Read Multihash where
readsPrec _ = readP_to_S parseMultihash
parseMultihash :: ReadP Multihash
parseMultihash = do
base58 <- munch1 (`elem` ("123456789ABCDEFGHJKLMNPQRSTUVWXYZabcdefghijkmnopqrstuvwxyz" :: String)) -- base58 bitcoin alphabet
either fail (return . Multihash) . MB.decode MB.Base58 . fromString $ base58
instance FromJSON Multihash where
parseJSON (JSON.String s) = return $ read . unpack $ s
parseJSON _ = fail "Expected a Multihash String"
type Key = ByteString
type Data = ByteString
data Template = Unixfs | None deriving (Show, Eq)
data FileHash = FileHash {
fileName :: FilePath,
fileHash :: Multihash
} deriving (Generic, Show, Eq)
instance FromJSON FileHash where
parseJSON (JSON.Object o) = FileHash
<$> o .: "Name"
<*> o .: "Hash"
parseJSON _ = fail "Expected a FileHash"
data Link = Link {
hash :: Maybe Multihash,
name :: Maybe FilePath,
size :: Maybe Int
} deriving (Generic, Eq, Show)
instance FromJSON Link where
parseJSON (JSON.Object o) = Link
<$> o .:? "Hash"
<*> o .:? "Name"
<*> o .:? "Size"
parseJSON _ = fail "Expected a Link"
data Node = Node {
links :: [Link],
payload :: Maybe Data
} deriving (Generic, Show, Eq)
instance FromJSON Node where
parseJSON (JSON.Object o) = Node
<$> o .: "Links"
<*> o .:? "Data"
parseJSON _ = fail "Expected a Node"
data Object = Object {
objectHash :: Multihash,
objectPayload :: Data,
objectLinks :: [(String, Object)]
} deriving (Generic, Show, Eq)
data ID = ID {
idHash :: Multihash,
publicKey :: Key,
addresses :: [Multiaddr], -- TODO replace with multiaddresses ?
agentVersion :: String,
protocolVersion :: String
} deriving (Generic, Show, Eq)
instance FromJSON ID where
parseJSON (JSON.Object o) = ID
<$> o .: "ID"
<*> o .: "PublicKey"
<*> (map read <$> o .: "Addresses")
<*> o .: "AgentVersion"
<*> o .: "ProtocolVersion"
parseJSON _ = fail "Expected an ID"
instance FromJSON ByteString where
parseJSON = (fromString <$>) . parseJSON
| basile-henry/hs-ipfs-api | src/Network/IPFS/Types.hs | mit | 4,247 | 0 | 14 | 1,313 | 1,150 | 657 | 493 | 111 | 1 |
{-# LANGUAGE FlexibleInstances #-}
{-# LANGUAGE OverloadedStrings #-}
{-# LANGUAGE DataKinds #-}
{-# LANGUAGE TypeSynonymInstances #-}
{-# OPTIONS_GHC -fno-warn-orphans #-}
module IHaskell.Display.Widgets.Output
( -- * The Output Widget
OutputWidget
-- * Constructor
, mkOutput
-- * Using the output widget
, appendStdout
, appendStderr
, appendDisplay
, clearOutput
, clearOutput_
, replaceOutput
) where
-- To keep `cabal repl` happy when running from the ihaskell repo
import Prelude
import Data.Aeson
import Data.IORef (newIORef)
import Data.Text
import Data.Vinyl (Rec(..), (<+>))
import IHaskell.Display
import IHaskell.Eval.Widgets
import IHaskell.IPython.Types (StreamType(..))
import IHaskell.IPython.Message.UUID as U
import IHaskell.Display.Widgets.Types
import IHaskell.Display.Widgets.Common
import IHaskell.Display.Widgets.Layout.LayoutWidget
-- | An 'OutputWidget' represents a Output widget from IPython.html.widgets.
type OutputWidget = IPythonWidget 'OutputType
-- | Create a new output widget
mkOutput :: IO OutputWidget
mkOutput = do
-- Default properties, with a random uuid
wid <- U.random
layout <- mkLayout
let domAttrs = defaultDOMWidget "OutputView" "OutputModel" layout
outAttrs = (ViewModule =:! "@jupyter-widgets/output")
:& (ModelModule =:! "@jupyter-widgets/output")
:& (ViewModuleVersion =:! "1.0.0")
:& (ModelModuleVersion =:! "1.0.0")
:& (MsgID =:: "")
:& (Outputs =:: [])
:& RNil
widgetState = WidgetState $ domAttrs <+> outAttrs
stateIO <- newIORef widgetState
let widget = IPythonWidget wid stateIO
-- Open a comm for this widget, and store it in the kernel state
widgetSendOpen widget $ toJSON widgetState
-- Return the image widget
return widget
-- | Appends the Text to the given Stream
appendStd :: StreamType -> OutputWidget -> Text -> IO ()
appendStd n out t = do
getField out Outputs >>= setField out Outputs . updateOutputs
where updateOutputs :: [OutputMsg] -> [OutputMsg]
updateOutputs = (++[OutputStream n t])
-- | Appends text to the stdout of an output widget
appendStdout :: OutputWidget -> Text -> IO ()
appendStdout = appendStd Stdout
-- | Appends text to the stderr of an output widget
appendStderr :: OutputWidget -> Text -> IO ()
appendStderr = appendStd Stderr
-- | Clears the output widget
clearOutput' :: OutputWidget -> IO ()
clearOutput' w = do
_ <- setField w Outputs []
_ <- setField w MsgID ""
return ()
-- | Appends anything displayable to an output widget
appendDisplay :: IHaskellDisplay a => OutputWidget -> a -> IO ()
appendDisplay o d = do
outputs <- getField o Outputs
disp <- display d
_ <- setField o Outputs $ outputs ++ [OutputData disp]
return ()
-- | Clear the output widget immediately
clearOutput :: OutputWidget -> IO ()
clearOutput widget = widgetClearOutput False >> clearOutput' widget
-- | Clear the output widget on next append
clearOutput_ :: OutputWidget -> IO ()
clearOutput_ widget = widgetClearOutput True >> clearOutput' widget
-- | Replace the currently displayed output for output widget
replaceOutput :: IHaskellDisplay a => OutputWidget -> a -> IO ()
replaceOutput widget d = do
disp <- display d
setField widget Outputs [OutputData disp]
instance IHaskellWidget OutputWidget where
getCommUUID = uuid
| gibiansky/IHaskell | ihaskell-display/ihaskell-widgets/src/IHaskell/Display/Widgets/Output.hs | mit | 3,538 | 0 | 17 | 810 | 770 | 408 | 362 | 75 | 1 |
{-# LANGUAGE ScopedTypeVariables #-}
module CNFList (tests) where
import BooleanLaws
import SimplifyLaws
import BooleanModelLaws
import FreeBoolean
import Test.Tasty
import Test.QuickCheck
import Control.Applicative
import Data.Function (on)
import Data.Algebra.Boolean.Negable (Negable, Neg(..))
import Data.Algebra.Boolean.CNF.List
instance Arbitrary a => Arbitrary (Neg a) where
arbitrary = oneof [ Pos <$> arbitrary, Pos <$> arbitrary ]
instance (Negable a, Arbitrary a) => Arbitrary (CNF a) where
arbitrary = fromFreeBoolean <$> arbitrary
tests :: TestTree
tests = testGroup "CNF list implementation"
[ monotoneLaws eq
, simplifyLaws (undefined :: CNF (Either Bool Bool))
, booleanModelLaws (undefined :: CNF (Either Bool Bool))
]
eq :: CNF (Neg Int) -> CNF (Neg Int) -> Bool
eq = (==) `on` toBool
| phadej/boolean-normal-forms | tests/CNFList.hs | mit | 824 | 0 | 11 | 131 | 266 | 150 | 116 | 23 | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.