Added tests of latex tokenizer.

This should help prevent regressions like #4159.
This commit is contained in:
John MacFarlane 2017-12-15 10:13:43 -08:00
parent 3a3d661408
commit 79c3f57c47

View file

@ -3,7 +3,11 @@ module Tests.Readers.LaTeX (tests) where
import Data.Text (Text)
import qualified Data.Text as T
import qualified Text.Pandoc.UTF8 as UTF8
import Text.Pandoc.Readers.LaTeX (tokenize, untokenize)
import Test.Tasty
import Test.Tasty.HUnit
import Test.Tasty.QuickCheck
import Tests.Helpers
import Text.Pandoc
import Text.Pandoc.Arbitrary ()
@ -22,8 +26,21 @@ simpleTable' :: [Alignment] -> [[Blocks]] -> Blocks
simpleTable' aligns = table "" (zip aligns (repeat 0.0))
(map (const mempty) aligns)
tokUntokRt :: String -> Bool
tokUntokRt s = untokenize (tokenize "random" t) == t
where t = T.pack s
tests :: [TestTree]
tests = [ testGroup "basic"
tests = [ testGroup "tokenization"
[ testCase "tokenizer round trip on test case" $ do
orig <- T.pack <$> UTF8.readFile "../test/latex-reader.latex"
let new = untokenize $ tokenize "../test/latex-reader.latex"
orig
assertEqual "untokenize . tokenize is identity" orig new
, testProperty "untokenize . tokenize is identity" tokUntokRt
]
, testGroup "basic"
[ "simple" =:
"word" =?> para "word"
, "space" =: