X-Git-Url: http://gitweb.michael.orlitzky.com/?a=blobdiff_plain;f=src%2FTSN%2FXML%2FNews.hs;h=10c51956fa20c23a16ff45b3469bc477aac078a7;hb=9b28d5ed0e1570f0890eae1ee01721688c20e266;hp=ea1378725e82e1cce381c83a9c8ea4b8c477ca4a;hpb=0ddc996d5622f138953ad027cc425713a276b46a;p=dead%2Fhtsn-import.git diff --git a/src/TSN/XML/News.hs b/src/TSN/XML/News.hs index ea13787..10c5195 100644 --- a/src/TSN/XML/News.hs +++ b/src/TSN/XML/News.hs @@ -3,22 +3,22 @@ {-# LANGUAGE GADTs #-} {-# LANGUAGE QuasiQuotes #-} {-# LANGUAGE RecordWildCards #-} -{-# LANGUAGE StandaloneDeriving #-} {-# LANGUAGE TemplateHaskell #-} {-# LANGUAGE TypeFamilies #-} --- | Parse TSN XML for the DTD "newsxml.dtd". Each document contains a --- root element \ that contains an entire news item. +-- | Parse TSN XML for the DTD \"newsxml.dtd\". Each document contains +-- a root element \ that contains an entire news item. -- module TSN.XML.News ( + dtd, + has_only_single_sms, pickle_message, -- * Tests news_tests, -- * WARNING: these are private but exported to silence warnings - News_NewsLocationConstructor(..), + News_LocationConstructor(..), News_NewsTeamConstructor(..), NewsConstructor(..), - NewsLocationConstructor(..), NewsTeamConstructor(..) ) where @@ -29,9 +29,15 @@ import Data.List.Utils ( join, split ) import Data.Tuple.Curry ( uncurryN ) import Data.Typeable ( Typeable ) import Database.Groundhog ( + countAll, + deleteAll, insert_, - migrate ) + migrate, + runMigration, + silentMigrationLogger ) import Database.Groundhog.Core ( DefaultKey ) +import Database.Groundhog.Generic ( runDbConn ) +import Database.Groundhog.Sqlite ( withSqliteConn ) import Database.Groundhog.TH ( defaultCodegenConfig, groundhog, @@ -40,6 +46,16 @@ import Test.Tasty ( TestTree, testGroup ) import Test.Tasty.HUnit ( (@?=), testCase ) import Text.XML.HXT.Core ( PU, + XmlTree, + (/>), + (>>>), + addNav, + descendantAxis, + filterAxis, + followingSiblingAxis, + hasName, + remNav, + runLA, xp13Tuple, xpAttr, xpElem, @@ -48,17 +64,30 @@ import Text.XML.HXT.Core ( xpOption, xpPair, xpText, - xpTriple, xpWrap ) -- Local imports. import TSN.Codegen ( tsn_codegen_config, tsn_db_field_namer ) -- Used in a test +import TSN.Database ( insert_or_select ) import TSN.DbImport ( DbImport(..), ImportResult(..), run_dbmigrate ) import TSN.Picklers ( xp_time_stamp ) +import TSN.Location ( Location(..), pickle_location ) import TSN.XmlImport ( XmlImport(..) ) -import Xml ( FromXml(..), ToDb(..), pickle_unpickle, unpickleable ) +import Xml ( + FromXml(..), + ToDb(..), + pickle_unpickle, + unpickleable, + unsafe_read_document, + unsafe_unpickle ) + + +-- | The DTD to which this module corresponds. Used to invoke dbimport. +-- +dtd :: String +dtd = "newsxml.dtd" -- @@ -67,10 +96,11 @@ import Xml ( FromXml(..), ToDb(..), pickle_unpickle, unpickleable ) -- * News/Message --- | The msg_id child of contains an event_id attribute; we --- embed it into the 'News' type. We (pointlessly) use the "db_" --- prefix here so that the two names don't collide on "id" when +-- | The msg_id child of \ contains an event_id attribute; we +-- embed it into the 'News' type. We (pointlessly) use the \"db_\" +-- prefix here so that the two names don't collide on \"id\" when -- Groundhog is creating its fields using our field namer. +-- data MsgId = MsgId { db_msg_id :: Int, @@ -78,7 +108,7 @@ data MsgId = deriving (Data, Eq, Show, Typeable) --- | The XML representation of a news item (message). +-- | The XML representation of a news item (\). -- data Message = Message { @@ -89,7 +119,7 @@ data Message = xml_sport :: String, xml_url :: Maybe String, xml_teams :: [NewsTeam], - xml_locations :: [NewsLocation], + xml_locations :: [Location], xml_sms :: String, xml_editor :: Maybe String, xml_text :: Maybe String, -- Text and continue seem to show up in pairs, @@ -118,6 +148,7 @@ data News = instance ToDb Message where + -- | The database representation of 'Message' is 'News'. type Db Message = News -- | Convert the XML representation 'Message' to the database @@ -137,36 +168,26 @@ instance FromXml Message where db_continue = xml_continue, db_time_stamp = xml_time_stamp } --- | This lets us call 'insert_xml' on a 'Message'. +-- | This lets us insert the XML representation 'Message' directly. -- instance XmlImport Message -- * NewsTeam --- | The database type for teams as they show up in the news. +-- | The database/XML type for teams as they show up in the news. We +-- can't reuse the representation from "TSN.Team" because they +-- require a team id. We wouldn't want to make the team ID optional +-- and then insert a team with no id, only to find the same team +-- later with an id and be unable to update the record. (We could +-- add the update logic, but it would be more trouble than it's +-- worth.) -- data NewsTeam = NewsTeam { team_name :: String } deriving (Eq, Show) -instance ToDb NewsTeam where - -- | The database representaion of a 'NewsTeam' is itself. - type Db NewsTeam = NewsTeam - --- | This is needed to define the XmlImport instance for NewsTeam. --- -instance FromXml NewsTeam where - -- | How to we get a 'NewsTeam' from itself? - from_xml = id - --- | Allow us to call 'insert_xml' on the XML representation of --- NewsTeams. --- -instance XmlImport NewsTeam - - -- * News_NewsTeam @@ -180,81 +201,82 @@ data News_NewsTeam = News_NewsTeam (DefaultKey NewsTeam) --- * NewsLocation +-- * News_Location --- | The database type for locations as they show up in the news. --- -data NewsLocation = - NewsLocation { - city :: Maybe String, - state :: Maybe String, - country :: String } - deriving (Eq, Show) - -instance ToDb NewsLocation where - -- | The database representation of a 'NewsLocation' is itself. - type Db NewsLocation = NewsLocation - --- | This is needed to define the XmlImport instance for NewsLocation. +-- | Mapping between 'News' records and 'Location' records in the +-- database. We don't name the fields because we don't use the names +-- explicitly; that means we have to give them nice database names +-- via groundhog. -- -instance FromXml NewsLocation where - -- | How to we get a 'NewsLocation' from itself? - from_xml = id +data News_Location = News_Location + (DefaultKey News) + (DefaultKey Location) --- | Allow us to call 'insert_xml' on the XML representation of --- NewsLocations. --- -instance XmlImport NewsLocation --- * News_NewsLocation --- | Mapping between News records and NewsLocation records in the --- database. We don't name the fields because we don't use the names --- explicitly; that means we have to give them nice database names --- via groundhog. +-- | Some newsxml documents contain two \ elements in a row, +-- violating the DTD. The second one has always been empty, but it's +-- irrelevant: we can't parse these, and would like to detect them +-- in order to report the fact that the busted document is +-- unsupported. +-- +-- This function detects whether two \ elements appear in a +-- row, as siblings. -- -data News_NewsLocation = News_NewsLocation - (DefaultKey News) - (DefaultKey NewsLocation) +has_only_single_sms :: XmlTree -> Bool +has_only_single_sms xmltree = + case elements of + [] -> True + _ -> False + where + parse :: XmlTree -> [XmlTree] + parse = runLA $ hasName "/" + /> hasName "message" + >>> addNav + >>> descendantAxis + >>> filterAxis (hasName "SMS") + >>> followingSiblingAxis + >>> remNav + >>> hasName "SMS" + elements = parse xmltree -- --- Database code +-- * Database code -- -- | Define 'dbmigrate' and 'dbimport' for 'Message's. The import is -- slightly non-generic because of our 'News_NewsTeam' and --- 'News_NewsLocation' join tables. +-- 'News_Location' join tables. -- instance DbImport Message where dbmigrate _ = run_dbmigrate $ do - migrate (undefined :: NewsTeam) - migrate (undefined :: NewsLocation) + migrate (undefined :: Location) migrate (undefined :: News) + migrate (undefined :: NewsTeam) migrate (undefined :: News_NewsTeam) - migrate (undefined :: News_NewsLocation) + migrate (undefined :: News_Location) dbimport message = do -- Insert the message and acquire its primary key (unique ID) news_id <- insert_xml message - -- And insert each one into its own table. We use insert_xml_or_select - -- because we know that most teams will already exist, and we - -- want to get back the id for the existing team when - -- there's a collision. - nt_ids <- mapM insert_xml_or_select (xml_teams message) + -- Now insert the teams. We use insert_or_select because we know + -- that most teams will already exist, and we want to get back the + -- id for the existing team when there's a collision. + nt_ids <- mapM insert_or_select (xml_teams message) -- Now that the teams have been inserted, create -- news__news_team records mapping beween the two. let news_news_teams = map (News_NewsTeam news_id) nt_ids mapM_ insert_ news_news_teams - -- Do all of that over again for the NewsLocations. - loc_ids <- mapM insert_xml_or_select (xml_locations message) - let news_news_locations = map (News_NewsLocation news_id) loc_ids + -- Do all of that over again for the Locations. + loc_ids <- mapM insert_or_select (xml_locations message) + let news_news_locations = map (News_Location news_id) loc_ids mapM_ insert_ news_news_locations return ImportSucceeded @@ -273,15 +295,6 @@ mkPersist defaultCodegenConfig [groundhog| type: constraint fields: [team_name] -- entity: NewsLocation - dbName: news_locations - constructors: - - name: NewsLocation - uniques: - - name: unique_news_location - type: constraint - fields: [city, state, country] - |] @@ -289,7 +302,6 @@ mkPersist defaultCodegenConfig [groundhog| -- use our own codegen to peel those off before naming the columns. mkPersist tsn_codegen_config [groundhog| - entity: News - dbName: news constructors: - name: News uniques: @@ -324,21 +336,22 @@ mkPersist tsn_codegen_config [groundhog| reference: onDelete: cascade -- entity: News_NewsLocation - dbName: news__news_locations +- entity: News_Location + dbName: news__locations constructors: - - name: News_NewsLocation + - name: News_Location fields: - - name: news_NewsLocation0 # Default created by mkNormalFieldName + - name: news_Location0 # Default created by mkNormalFieldName dbName: news_id reference: onDelete: cascade - - name: news_NewsLocation1 # Default created by mkNormalFieldName - dbName: news_locations_id + - name: news_Location1 # Default created by mkNormalFieldName + dbName: locations_id reference: onDelete: cascade |] + -- -- XML Picklers -- @@ -369,20 +382,6 @@ pickle_msg_id = to_tuple m = (db_msg_id m, db_event_id m) --- | Convert a 'NewsLocation' to/from XML. --- -pickle_location :: PU NewsLocation -pickle_location = - xpElem "location" $ - xpWrap (from_tuple, to_tuple) $ - xpTriple (xpOption (xpElem "city" xpText)) - (xpOption (xpElem "state" xpText)) - (xpElem "country" xpText) - where - from_tuple = - uncurryN NewsLocation - to_tuple l = (city l, state l, country l) - -- | Convert a 'Message' to/from XML. -- @@ -447,8 +446,10 @@ news_tests = testGroup "News tests" [ test_news_fields_have_correct_names, + test_on_delete_cascade, test_pickle_of_unpickle_is_identity, - test_unpickle_succeeds ] + test_unpickle_succeeds, + test_sms_detected_correctly ] -- | Make sure our codegen is producing the correct database names. @@ -469,7 +470,14 @@ test_news_fields_have_correct_names = map (\x -> tsn_db_field_namer "herp" "derp" 8675309 x 90210) field_names actual :: [String] - actual = ["mid", "sport", "url", "sms", "editor", "text", "continue"] + actual = ["xml_file_id", + "mid", + "sport", + "url", + "sms", + "editor", + "text", + "continue"] check (x,y) = (x @?= y) @@ -505,3 +513,57 @@ test_unpickle_succeeds = testGroup "unpickle tests" actual <- unpickleable path pickle_message let expected = True actual @?= expected + + +-- | Make sure everything gets deleted when we delete the top-level +-- record. +-- +test_on_delete_cascade :: TestTree +test_on_delete_cascade = testGroup "cascading delete tests" + [ check "deleting news deletes its children" + "test/xml/newsxml.xml" + 4 -- 2 news_teams and 2 news_locations that should remain. + ] + where + check desc path expected = testCase desc $ do + news <- unsafe_unpickle path pickle_message + let a = undefined :: Location + let b = undefined :: News + let c = undefined :: NewsTeam + let d = undefined :: News_NewsTeam + let e = undefined :: News_Location + actual <- withSqliteConn ":memory:" $ runDbConn $ do + runMigration silentMigrationLogger $ do + migrate a + migrate b + migrate c + migrate d + migrate e + _ <- dbimport news + deleteAll b + count_a <- countAll a + count_b <- countAll b + count_c <- countAll c + count_d <- countAll d + count_e <- countAll e + return $ count_a + count_b + count_c + count_d + count_e + actual @?= expected + + +-- | We want to make sure the single-SMS documents and the multi-SMS +-- documents are identified correctly. +-- +test_sms_detected_correctly :: TestTree +test_sms_detected_correctly = + testGroup "newsxml SMS count determined correctly" + [ check "test/xml/newsxml.xml" + "single SMS detected correctly" + True, + check "test/xml/newsxml-multiple-sms.xml" + "multiple SMS detected correctly" + False ] + where + check path desc expected = testCase desc $ do + xmltree <- unsafe_read_document path + let actual = has_only_single_sms xmltree + actual @?= expected