{-| Module : Frames.SQL.Beam.Postgres Description : Re-exports the underlying modules. Contains usage examples. Copyright : (c) Gagandeep Bhatia, 2018 License : BSD3 Stability : experimental This library provides a way to access Postgres tables as data frames by providing helpers for generating types (at compile time) corresponding to a database schema ("Frames.SQL.Beam.Postgres.BeamSchemaGen") and canned queries to execute against a database ("Frames.SQL.Beam.Postgres.Query") instance. Additionally, provides utilities to convert plain Haskell records (i.e. the format of query results) to vinyl records (upon which the Frames library is based) ("Frames.SQL.Beam.Postgres.Vinylize"). Can be used for interactive exploration by loading all data in-memory at once (and converting to a data frame), and also in a constant memory streaming mode ("Frames.SQL.Beam.Postgres.Streaming"). -} module Frames.SQL.Beam.Postgres ( -- * Usage Example -- ** A. Interactive Workflow Steps -- $interactive -- ** B. Streaming Workflow Steps -- $streaming -- * A Note on 'Filter Lambdas' -- $filter-lambda-note -- * Re-exports of the underlying modules module Frames.SQL.Beam.Postgres.Vinylize , module Frames.SQL.Beam.Postgres.BeamSchemaGen , module Frames.SQL.Beam.Postgres.Query , module Frames.SQL.Beam.Postgres.Streaming -- * Re-exports of @beam-core@ and @beam-postgres@ , module Database.Beam , module Database.Beam.Postgres -- * Postgres Column Types , ByteString , Text , UUID , Scientific , UTCTime , LocalTimestamp , UTCTimestamp , ZonedTimestamp , LocalTime , TimeOfDay , Date , Day -- * Re-exports for the @deriveGeneric@ plus @deriveVinyl@ combination , module Data.Coerce , module Generics.SOP , deriveGeneric ) where import Frames.SQL.Beam.Postgres.BeamSchemaGen import Frames.SQL.Beam.Postgres.Query import Frames.SQL.Beam.Postgres.Streaming import Frames.SQL.Beam.Postgres.Vinylize import Database.Beam import Database.Beam.Postgres import Data.ByteString (ByteString) import Data.Scientific (Scientific) import Data.Text (Text) import Data.Time (Day, LocalTime, TimeOfDay, UTCTime) import Data.UUID.Types (UUID) import Database.PostgreSQL.Simple.Time (Date, LocalTimestamp, UTCTimestamp, ZonedTimestamp) import Data.Coerce import Generics.SOP hiding (Generic, fieldName) import Generics.SOP.TH (deriveGeneric) {- $streaming Once you're done working with a small subset of data, and would like to scale up your analysis by looking at a larger-subset-of/complete data, then it's time to look at writing your own `conduit` to process incoming rows from the DB. @1 - 4:@ Same as 'Interactive Workflow Steps' @5.@ Writing your own streaming pipeline: Consider the following: @ streamRows :: IO () streamRows = do res <- 'withConnection' connString $ 'streamingSelectAllPipeline'' _cart_users db 1000 (\c -> (_cart_usersFirst_name c) \`like_\` \"J%\") $ (CL.map (\record -> F.rcast @["_cart_usersEmail" F.:-> Text, "_cart_usersIs_member" F.:-> Bool] record)) mapM_ print res @ In the above, we select all rows from the specified table that match a certain pattern (@\"J%\"@), then the function `streamingSelectAllPipeline'` converts the query results to vinyl records inside a @conduit@ and sends it downstream, where we can operate on its output. Here, specifically, we do a column subset of the output using @rcast@, and @CL.map@ applies @rcast@ to every incoming row and sends it downstream, where the result gets returned. We then print the list of `vinyl` records. In order to write your own conduit, all you need to know is that internally the conduit flow is as follows: @ (\c -> runConduit $ c .| CL.map 'createRecId' .| recordProcessorConduit .| CL.take nrows) @ In the above, you supply the `recordProcessorConduit` to the `streamingSelectAllPipeline'` function which takes a @vinyl@ record as input and sends it downstream to the @CL.take@. Note that in all functions in the "Frames.SQL.Beam.Postgres.Streaming" module, you need to specify the number of rows you want to return (this is an upper bound of sorts, the actual number of rows returned depends on the amount of data present in your database). -} {- $interactive @1.@ Bootstrap database schema: In a new project, assume a file @Example.hs@ is present in the @src@ directory with the code below. You may of course change the string passed to `genBeamSchema` to match your database instance of interest. @ -- Example.hs -- Extensions elided module Example where import qualified Data.Conduit.List as CL import qualified Data.Vinyl.Functor as VF import qualified Frames as F import Frames.SQL.Beam.Postgres $('genBeamSchema' "host=localhost dbname=shoppingcart1") @ @2.@ Next, execute @stack build@ or @stack ghci@. This compilation step, if completed without any errors, will establish a connection to your database instance of interest, read its schema, generate corresponding Haskell types and put them in a module named @NewBeamSchema@ in your @src@ directory (the file creation step is also part of the compilation process). @3.@ Assuming step 2 worked fine for you and you were using the test DB-dump from the @data@ folder in the repo, you should now have a module with code matching that in the @test/NewBeamSchema.hs@ file of this repository. In case you used some other database instance of your own, your generated module would look different. Import this module into @Example@: @ -- Example.hs -- Extensions elided module Example where import qualified Data.Conduit.List as CL import qualified Data.Vinyl.Functor as VF import qualified Frames as F import Frames.SQL.Beam.Postgres import NewBeamSchema $('genBeamSchema' "host=localhost dbname=shoppingcart1") @ @4.@ Let's assume the table of interest is @Cart_usersT@. We want to pull rows from this table into a data frame to explore it interactively from @ghci@. Note that @beam@ query results are lists of plain Haskell records whereas "Frames" requires a list of @vinyl@ records. In order to make this conversion, we add the following two invokations of code-generating (Template-Haskell) functions to @Example@: @ -- Example.hs -- rest of the module elided import NewBeamSchema $('genBeamSchema' "host=localhost dbname=shoppingcart1") 'deriveGeneric' ''Cart_usersT 'deriveVinyl' ''Cart_usersT @ ...and build your project. This will add some additional code into the @Example@ module. You can inspect this code by adding the appropriate compiler flags to your @.cabal@ file. @5.@ Querying the DB: In this step we will execute a 'SELECT * FROM tbl WHERE...' query and convert the results to a data frame. Note that the table declaration (@_cart_users@) and the database declaration (@db@) are exported by the @NewBeamSchema@ module. More importantly, these declarations are autogenerated at compile time, so in case new tables are added, the corresponding declarations are automatically available for use. @ -- Example.hs connString :: ByteString connString = "host=localhost dbname=shoppingcart1" -- selects 'n' rows from the specified table in the db. loadRows1 :: Int -> IO [(Cart_usersT Identity)] loadRows1 n = 'withConnection' connString $ 'bulkSelectAllRows' _cart_users db n loadRows2 :: Int -> IO [(Cart_usersT Identity)] loadRows2 n = 'withConnection' connString $ 'bulkSelectAllRowsWhere' _cart_users db n (\c -> (_cart_usersFirst_name c) \`like_\` \"J%\") @ Notice the lambda passed to `bulkSelectAllRowsWhere` in `loadRows2`. This is a 'filter lambda' that forms the `WHERE ...` part of the SQL query and is executed at the DB-level. We will see how to create our own 'filter lambdas' in another section below. For now, if we were to enter @ghci@ by executing @stack ghci@ after adding the above code: @ ghci>res1 <- loadRows1 5 ghci>:t res1 res1 :: [Cart_usersT Identity] ghci>:t (map createRecId res1) (map createRecId res1) :: [F.Rec VF.Identity '["_cart_usersEmail" F.:-> Text, "_cart_usersFirst_name" F.:-> Text, "_cart_usersLast_name" F.:-> Text, "_cart_usersIs_member" F.:-> Bool, "_cart_usersDays_in_queue" F.:-> Int]] ghci>:t (F.toFrame $ map createRecId res1) (F.toFrame $ map createRecId res1) :: F.Frame (F.Record '["_cart_usersEmail" F.:-> Text, "_cart_usersFirst_name" F.:-> Text, "_cart_usersLast_name" F.:-> Text, "_cart_usersIs_member" F.:-> Bool, "_cart_usersDays_in_queue" F.:-> Int]) ghci>myFrame = F.toFrame $ map createRecId res1 ghci>:set -XTypeApplications ghci>:set -XTypeOperators ghci>:set -XDataKinds ghci>miniFrame = fmap (F.rcast @'["_cart_usersEmail" F.:-> Text, "_cart_usersDays_in_queue" F.:-> Int]) myFrame ghci>mapM_ print miniFrame {_cart_usersEmail :-> "james@example.com", _cart_usersDays_in_queue :-> 1} {_cart_usersEmail :-> "betty@example.com", _cart_usersDays_in_queue :-> 42} {_cart_usersEmail :-> "james@pallo.com", _cart_usersDays_in_queue :-> 1} {_cart_usersEmail :-> "betty@sims.com", _cart_usersDays_in_queue :-> 42} {_cart_usersEmail :-> "james@oreily.com", _cart_usersDays_in_queue :-> 1} @ We could have used `loadRows2` in place of `loadRows1` in order to have the @WHERE ...@ clause executed at the DB-level. Note that in the above, once the query results are converted to a data frame, you're free to play with the frame in anyway, just like you would for a data frame created from a CSV. -} {- $filter-lambda-note A Note on 'Canned Queries' and 'Filter Lambdas': There are three things needed to execute a canned query ('SELECT * FROM tbl WHERE ...'): * @PostgresTable a b@: auto generated by BeamSchemaGen module * @PostgresDB b@: auto generated by BeamSchemaGen module * @PostgresFilterLambda a s@: The @WHERE...@ clause. All filter lambdas are of the form: @ (\tbl -> (_fieldName tbl) `op` constant) @ or @ (\tbl -> (_fieldName1 tbl) `op` (_fieldName2 tbl)) @ In the above `op` can be one of : [`==.`, `/=.`, `>.`, `<.`, `<=.`, `>=.`, `between_`, `like_`, `in_` ] (some of these are not be applicable to the second case). You may use `&&.` and `||.` to combine expressions inside the lambda. To see some actual examples of 'filter lambdas', check out @test/LibSpec.hs@ in this repository. -}